var/home/core/zuul-output/0000755000175000017500000000000015115047317014531 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015115060263015470 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005060642515115060254017704 0ustar rootrootDec 06 15:45:59 crc systemd[1]: Starting Kubernetes Kubelet... Dec 06 15:45:59 crc restorecon[4582]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:45:59 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 15:46:00 crc restorecon[4582]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 15:46:00 crc restorecon[4582]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 06 15:46:00 crc kubenswrapper[4813]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 15:46:00 crc kubenswrapper[4813]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 06 15:46:00 crc kubenswrapper[4813]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 15:46:00 crc kubenswrapper[4813]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 15:46:00 crc kubenswrapper[4813]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 06 15:46:00 crc kubenswrapper[4813]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.298753 4813 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305177 4813 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305210 4813 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305215 4813 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305219 4813 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305223 4813 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305227 4813 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305232 4813 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305238 4813 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305243 4813 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305248 4813 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305252 4813 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305255 4813 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305278 4813 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305282 4813 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305285 4813 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305289 4813 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305292 4813 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305297 4813 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305303 4813 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305309 4813 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305314 4813 feature_gate.go:330] unrecognized feature gate: Example Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305319 4813 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305323 4813 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305328 4813 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305333 4813 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305339 4813 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305343 4813 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305347 4813 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305351 4813 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305355 4813 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305359 4813 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305362 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305366 4813 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305369 4813 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305381 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305385 4813 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305388 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305392 4813 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305395 4813 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305399 4813 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305403 4813 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305406 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305410 4813 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305413 4813 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305417 4813 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305420 4813 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305424 4813 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305430 4813 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305435 4813 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305439 4813 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305443 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305448 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305451 4813 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305455 4813 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305459 4813 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305462 4813 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305467 4813 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305470 4813 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305473 4813 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305477 4813 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305480 4813 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305484 4813 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305487 4813 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305491 4813 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305494 4813 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305497 4813 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305500 4813 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305505 4813 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305509 4813 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305513 4813 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.305516 4813 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305613 4813 flags.go:64] FLAG: --address="0.0.0.0" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305623 4813 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305629 4813 flags.go:64] FLAG: --anonymous-auth="true" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305634 4813 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305640 4813 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305645 4813 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305651 4813 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305656 4813 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305661 4813 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305666 4813 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305671 4813 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305682 4813 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305686 4813 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305690 4813 flags.go:64] FLAG: --cgroup-root="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305695 4813 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305699 4813 flags.go:64] FLAG: --client-ca-file="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305703 4813 flags.go:64] FLAG: --cloud-config="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305707 4813 flags.go:64] FLAG: --cloud-provider="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305711 4813 flags.go:64] FLAG: --cluster-dns="[]" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305716 4813 flags.go:64] FLAG: --cluster-domain="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305720 4813 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305725 4813 flags.go:64] FLAG: --config-dir="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305729 4813 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305733 4813 flags.go:64] FLAG: --container-log-max-files="5" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305739 4813 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305744 4813 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305748 4813 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305752 4813 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305757 4813 flags.go:64] FLAG: --contention-profiling="false" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305761 4813 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305765 4813 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305769 4813 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305773 4813 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305779 4813 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305783 4813 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305787 4813 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305791 4813 flags.go:64] FLAG: --enable-load-reader="false" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305796 4813 flags.go:64] FLAG: --enable-server="true" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305800 4813 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305806 4813 flags.go:64] FLAG: --event-burst="100" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305810 4813 flags.go:64] FLAG: --event-qps="50" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305814 4813 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305820 4813 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305824 4813 flags.go:64] FLAG: --eviction-hard="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305834 4813 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305839 4813 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305843 4813 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305847 4813 flags.go:64] FLAG: --eviction-soft="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305871 4813 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305876 4813 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305880 4813 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305884 4813 flags.go:64] FLAG: --experimental-mounter-path="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305888 4813 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305893 4813 flags.go:64] FLAG: --fail-swap-on="true" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305897 4813 flags.go:64] FLAG: --feature-gates="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305902 4813 flags.go:64] FLAG: --file-check-frequency="20s" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305906 4813 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305911 4813 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305915 4813 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305920 4813 flags.go:64] FLAG: --healthz-port="10248" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305924 4813 flags.go:64] FLAG: --help="false" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305928 4813 flags.go:64] FLAG: --hostname-override="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305932 4813 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305936 4813 flags.go:64] FLAG: --http-check-frequency="20s" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305940 4813 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305944 4813 flags.go:64] FLAG: --image-credential-provider-config="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305948 4813 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305952 4813 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305957 4813 flags.go:64] FLAG: --image-service-endpoint="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305961 4813 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305965 4813 flags.go:64] FLAG: --kube-api-burst="100" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305969 4813 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305973 4813 flags.go:64] FLAG: --kube-api-qps="50" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305978 4813 flags.go:64] FLAG: --kube-reserved="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305982 4813 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305986 4813 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305990 4813 flags.go:64] FLAG: --kubelet-cgroups="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305994 4813 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.305998 4813 flags.go:64] FLAG: --lock-file="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306002 4813 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306007 4813 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306011 4813 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306017 4813 flags.go:64] FLAG: --log-json-split-stream="false" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306022 4813 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306026 4813 flags.go:64] FLAG: --log-text-split-stream="false" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306031 4813 flags.go:64] FLAG: --logging-format="text" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306035 4813 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306039 4813 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306043 4813 flags.go:64] FLAG: --manifest-url="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306047 4813 flags.go:64] FLAG: --manifest-url-header="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306053 4813 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306057 4813 flags.go:64] FLAG: --max-open-files="1000000" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306062 4813 flags.go:64] FLAG: --max-pods="110" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306066 4813 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306070 4813 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306075 4813 flags.go:64] FLAG: --memory-manager-policy="None" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306079 4813 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306083 4813 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306087 4813 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306091 4813 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306102 4813 flags.go:64] FLAG: --node-status-max-images="50" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306107 4813 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306111 4813 flags.go:64] FLAG: --oom-score-adj="-999" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306116 4813 flags.go:64] FLAG: --pod-cidr="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306119 4813 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306126 4813 flags.go:64] FLAG: --pod-manifest-path="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306130 4813 flags.go:64] FLAG: --pod-max-pids="-1" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306134 4813 flags.go:64] FLAG: --pods-per-core="0" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306139 4813 flags.go:64] FLAG: --port="10250" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306143 4813 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306147 4813 flags.go:64] FLAG: --provider-id="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306151 4813 flags.go:64] FLAG: --qos-reserved="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306155 4813 flags.go:64] FLAG: --read-only-port="10255" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306159 4813 flags.go:64] FLAG: --register-node="true" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306163 4813 flags.go:64] FLAG: --register-schedulable="true" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306168 4813 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306175 4813 flags.go:64] FLAG: --registry-burst="10" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306179 4813 flags.go:64] FLAG: --registry-qps="5" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306183 4813 flags.go:64] FLAG: --reserved-cpus="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306188 4813 flags.go:64] FLAG: --reserved-memory="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306194 4813 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306198 4813 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306203 4813 flags.go:64] FLAG: --rotate-certificates="false" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306207 4813 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306211 4813 flags.go:64] FLAG: --runonce="false" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306215 4813 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306219 4813 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306223 4813 flags.go:64] FLAG: --seccomp-default="false" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306228 4813 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306232 4813 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306237 4813 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306241 4813 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306245 4813 flags.go:64] FLAG: --storage-driver-password="root" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306249 4813 flags.go:64] FLAG: --storage-driver-secure="false" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306254 4813 flags.go:64] FLAG: --storage-driver-table="stats" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306270 4813 flags.go:64] FLAG: --storage-driver-user="root" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306275 4813 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306279 4813 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306284 4813 flags.go:64] FLAG: --system-cgroups="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306288 4813 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306295 4813 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306299 4813 flags.go:64] FLAG: --tls-cert-file="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306303 4813 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306308 4813 flags.go:64] FLAG: --tls-min-version="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306312 4813 flags.go:64] FLAG: --tls-private-key-file="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306316 4813 flags.go:64] FLAG: --topology-manager-policy="none" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306320 4813 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306325 4813 flags.go:64] FLAG: --topology-manager-scope="container" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306329 4813 flags.go:64] FLAG: --v="2" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306335 4813 flags.go:64] FLAG: --version="false" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306341 4813 flags.go:64] FLAG: --vmodule="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306346 4813 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306351 4813 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306455 4813 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306459 4813 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306467 4813 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306471 4813 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306475 4813 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306479 4813 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306483 4813 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306488 4813 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306491 4813 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306495 4813 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306499 4813 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306503 4813 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306506 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306511 4813 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306514 4813 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306518 4813 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306528 4813 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306531 4813 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306535 4813 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306539 4813 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306544 4813 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306547 4813 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306551 4813 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306554 4813 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306558 4813 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306562 4813 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306565 4813 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306568 4813 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306572 4813 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306575 4813 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306579 4813 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306582 4813 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306586 4813 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306589 4813 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306592 4813 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306596 4813 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306600 4813 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306604 4813 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306609 4813 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306614 4813 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306619 4813 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306622 4813 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306626 4813 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306629 4813 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306633 4813 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306637 4813 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306640 4813 feature_gate.go:330] unrecognized feature gate: Example Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306644 4813 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306649 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306652 4813 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306655 4813 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306659 4813 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306663 4813 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306666 4813 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306670 4813 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306673 4813 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306677 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306681 4813 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306684 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306689 4813 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306694 4813 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306698 4813 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306702 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306706 4813 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306710 4813 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306714 4813 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306717 4813 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306721 4813 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306725 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306729 4813 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.306732 4813 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.306924 4813 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.318868 4813 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.319304 4813 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319465 4813 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319484 4813 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319496 4813 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319509 4813 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319521 4813 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319532 4813 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319543 4813 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319554 4813 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319565 4813 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319580 4813 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319596 4813 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319607 4813 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319621 4813 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319636 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319647 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319661 4813 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319674 4813 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319685 4813 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319697 4813 feature_gate.go:330] unrecognized feature gate: Example Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319708 4813 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319718 4813 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319729 4813 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319739 4813 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319749 4813 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319759 4813 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319769 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319779 4813 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319790 4813 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319802 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319813 4813 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319822 4813 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319832 4813 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319842 4813 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319852 4813 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319865 4813 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319876 4813 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319886 4813 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319897 4813 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319907 4813 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319917 4813 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319927 4813 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319939 4813 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319949 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319959 4813 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319969 4813 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319980 4813 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.319990 4813 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320000 4813 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320010 4813 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320020 4813 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320029 4813 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320038 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320048 4813 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320057 4813 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320067 4813 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320076 4813 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320085 4813 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320095 4813 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320105 4813 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320116 4813 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320127 4813 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320138 4813 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320149 4813 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320159 4813 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320169 4813 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320179 4813 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320243 4813 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320281 4813 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320296 4813 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320311 4813 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320325 4813 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.320342 4813 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320652 4813 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320668 4813 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320682 4813 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320744 4813 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320758 4813 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320770 4813 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320782 4813 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320795 4813 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320806 4813 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320819 4813 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320831 4813 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320842 4813 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320855 4813 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320865 4813 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320875 4813 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320885 4813 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320896 4813 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320907 4813 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320917 4813 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320928 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320937 4813 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320948 4813 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320958 4813 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320969 4813 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320979 4813 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320989 4813 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.320999 4813 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321010 4813 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321025 4813 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321037 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321047 4813 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321057 4813 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321071 4813 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321084 4813 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321098 4813 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321111 4813 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321121 4813 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321133 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321144 4813 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321158 4813 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321171 4813 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321186 4813 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321198 4813 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321208 4813 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321220 4813 feature_gate.go:330] unrecognized feature gate: Example Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321232 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321242 4813 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321252 4813 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321293 4813 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321304 4813 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321314 4813 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321325 4813 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321335 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321348 4813 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321359 4813 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321369 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321379 4813 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321389 4813 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321399 4813 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321409 4813 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321420 4813 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321430 4813 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321439 4813 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321450 4813 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321460 4813 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321470 4813 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321480 4813 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321490 4813 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321500 4813 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321510 4813 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.321521 4813 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.321537 4813 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.321945 4813 server.go:940] "Client rotation is on, will bootstrap in background" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.326551 4813 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.326690 4813 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.327456 4813 server.go:997] "Starting client certificate rotation" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.327494 4813 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.327955 4813 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-19 20:09:02.551407358 +0000 UTC Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.328024 4813 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.333209 4813 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 06 15:46:00 crc kubenswrapper[4813]: E1206 15:46:00.333650 4813 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.337596 4813 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.356751 4813 log.go:25] "Validated CRI v1 runtime API" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.373801 4813 log.go:25] "Validated CRI v1 image API" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.376535 4813 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.379085 4813 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-06-15-41-05-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.379124 4813 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.391940 4813 manager.go:217] Machine: {Timestamp:2025-12-06 15:46:00.390809692 +0000 UTC m=+0.281689308 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2800000 MemoryCapacity:25199480832 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:38943010-e20a-4c38-8e12-0258df15005d BootID:1c54b507-6054-49c8-a12e-f79e691f7aaa Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039898624 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599742464 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:3076108 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599738368 Type:vfs Inodes:3076108 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:06:41:2c Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:06:41:2c Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:99:90:6e Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:7e:1d:d3 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:b0:1f:2f Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:90:4a:09 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:aa:12:3e:c5:59:86 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:f2:e7:51:f5:2e:31 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199480832 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.392159 4813 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.392357 4813 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.400827 4813 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.401033 4813 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.401074 4813 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.401317 4813 topology_manager.go:138] "Creating topology manager with none policy" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.401332 4813 container_manager_linux.go:303] "Creating device plugin manager" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.401596 4813 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.401666 4813 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.401988 4813 state_mem.go:36] "Initialized new in-memory state store" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.402091 4813 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.402939 4813 kubelet.go:418] "Attempting to sync node with API server" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.402967 4813 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.402993 4813 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.403009 4813 kubelet.go:324] "Adding apiserver pod source" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.403021 4813 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.405024 4813 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.405417 4813 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.405570 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Dec 06 15:46:00 crc kubenswrapper[4813]: E1206 15:46:00.405746 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.406301 4813 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.406486 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.407011 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.407039 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.407047 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.407056 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.407069 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.407079 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.407088 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.407102 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.407114 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.407123 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.407136 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 06 15:46:00 crc kubenswrapper[4813]: E1206 15:46:00.406581 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.419190 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.419441 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.420710 4813 server.go:1280] "Started kubelet" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.421150 4813 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.422104 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.421477 4813 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 06 15:46:00 crc systemd[1]: Started Kubernetes Kubelet. Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.423031 4813 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.423232 4813 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.423293 4813 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.423667 4813 server.go:460] "Adding debug handlers to kubelet server" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.423841 4813 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.426081 4813 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 06 15:46:00 crc kubenswrapper[4813]: E1206 15:46:00.425729 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 06 15:46:00 crc kubenswrapper[4813]: E1206 15:46:00.425793 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="200ms" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.426021 4813 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.425324 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Dec 06 15:46:00 crc kubenswrapper[4813]: E1206 15:46:00.426356 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.425768 4813 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 01:40:11.594787823 +0000 UTC Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.433412 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.433462 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.433475 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.433487 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.433500 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.433513 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.433524 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.433535 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.433548 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.433579 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.433591 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.433602 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.433613 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.433627 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.433639 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.433650 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.433684 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.433698 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.433710 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.433723 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.433737 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.433749 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: E1206 15:46:00.426824 4813 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.9:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187eaadabcaa3215 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 15:46:00.420676117 +0000 UTC m=+0.311555703,LastTimestamp:2025-12-06 15:46:00.420676117 +0000 UTC m=+0.311555703,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.433761 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.433833 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.433879 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.433905 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.433937 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.433963 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.433986 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434006 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434027 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434047 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434068 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434088 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434109 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434130 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434152 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434171 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434191 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434212 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434236 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434256 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434317 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434338 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434360 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434381 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434401 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434458 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434481 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434502 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434537 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434559 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434587 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434609 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434639 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434661 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434683 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434706 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434729 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434748 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434767 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434787 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434807 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434827 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434848 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434870 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434893 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434913 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434932 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434952 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434972 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.434993 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435016 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435036 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435056 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435076 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435095 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435115 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435135 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435158 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435180 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435200 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435219 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435240 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435290 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435312 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435335 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435367 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435388 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435410 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435431 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435452 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435471 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435491 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435512 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435531 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435550 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435570 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435589 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435609 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435629 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435665 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435685 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435703 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435730 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435752 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435774 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435796 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435817 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435838 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435858 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435880 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435902 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435923 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435942 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435963 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.435981 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436002 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436021 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436042 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436062 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436083 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436101 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436120 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436140 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436158 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436178 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436197 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436220 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436240 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436289 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436311 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436331 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436351 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436370 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436390 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436409 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436430 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436449 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436467 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436487 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436508 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436526 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436546 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436565 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436585 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436604 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436631 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436651 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436670 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436689 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436708 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436728 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436747 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436765 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436783 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436803 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436822 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436841 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436860 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436879 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436897 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436916 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436936 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436956 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436975 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.436995 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.437013 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.437034 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.437054 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.437073 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.437093 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.437113 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.437460 4813 factory.go:153] Registering CRI-O factory Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.437492 4813 factory.go:221] Registration of the crio container factory successfully Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.437624 4813 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.437642 4813 factory.go:55] Registering systemd factory Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.437656 4813 factory.go:221] Registration of the systemd container factory successfully Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.437691 4813 factory.go:103] Registering Raw factory Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.437717 4813 manager.go:1196] Started watching for new ooms in manager Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.438733 4813 manager.go:319] Starting recovery of all containers Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.442230 4813 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.442302 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.442321 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.442336 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.442352 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.442368 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.442382 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.442397 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.442411 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.442478 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.442493 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.442855 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.442884 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.442901 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.442914 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.442928 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.442941 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.442955 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.442970 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.442983 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.442997 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.443011 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.443025 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.443039 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.443052 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.443067 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.443078 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.443092 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.443105 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.443121 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.443135 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.443149 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.443164 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.443179 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.443195 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.443208 4813 reconstruct.go:97] "Volume reconstruction finished" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.443217 4813 reconciler.go:26] "Reconciler: start to sync state" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.472349 4813 manager.go:324] Recovery completed Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.482944 4813 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.485476 4813 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.485517 4813 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.485542 4813 kubelet.go:2335] "Starting kubelet main sync loop" Dec 06 15:46:00 crc kubenswrapper[4813]: E1206 15:46:00.485593 4813 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.485793 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.486770 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Dec 06 15:46:00 crc kubenswrapper[4813]: E1206 15:46:00.486818 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.487715 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.487748 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.487760 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.488811 4813 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.488828 4813 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.488851 4813 state_mem.go:36] "Initialized new in-memory state store" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.499086 4813 policy_none.go:49] "None policy: Start" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.500484 4813 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.500630 4813 state_mem.go:35] "Initializing new in-memory state store" Dec 06 15:46:00 crc kubenswrapper[4813]: E1206 15:46:00.527193 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.551574 4813 manager.go:334] "Starting Device Plugin manager" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.551625 4813 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.551638 4813 server.go:79] "Starting device plugin registration server" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.551968 4813 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.552000 4813 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.552213 4813 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.552294 4813 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.552302 4813 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 06 15:46:00 crc kubenswrapper[4813]: E1206 15:46:00.561222 4813 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.586657 4813 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.586802 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.588195 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.588227 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.588236 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.588399 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.588980 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.589006 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.589711 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.589730 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.589738 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.590146 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.590164 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.590172 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.590274 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.590588 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.590619 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.591436 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.591458 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.591466 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.591543 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.591590 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.591623 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.591679 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.591830 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.591856 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.592529 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.592551 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.592560 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.592637 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.592725 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.592754 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.592765 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.592808 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.592853 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.593741 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.593768 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.593779 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.593822 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.593851 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.593862 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.594050 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.594080 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.594743 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.594777 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.594809 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:00 crc kubenswrapper[4813]: E1206 15:46:00.627107 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="400ms" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.652174 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.652766 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.652804 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.652823 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.652842 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.652856 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.652928 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.652977 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.653005 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.653027 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.653052 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.653069 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.653090 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.653147 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.653183 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.653203 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.653501 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.653546 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.653563 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.653597 4813 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 15:46:00 crc kubenswrapper[4813]: E1206 15:46:00.654061 4813 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.9:6443: connect: connection refused" node="crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.754727 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.754806 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.754840 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.754869 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.754898 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.754925 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.754952 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.754978 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.755005 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.755033 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.755060 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.755086 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.755115 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.755142 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.755170 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.755499 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.755538 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.755577 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.755641 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.755597 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.755638 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.755651 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.755698 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.755737 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.755748 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.755754 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.755770 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.755799 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.755800 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.755819 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.855245 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.857191 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.857238 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.857251 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.857301 4813 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 15:46:00 crc kubenswrapper[4813]: E1206 15:46:00.857706 4813 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.9:6443: connect: connection refused" node="crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.930778 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.953553 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.962700 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-26c1cead14106af47b3089b4c2053d618b3045d75fc3d1e58515b0c88ec6c139 WatchSource:0}: Error finding container 26c1cead14106af47b3089b4c2053d618b3045d75fc3d1e58515b0c88ec6c139: Status 404 returned error can't find the container with id 26c1cead14106af47b3089b4c2053d618b3045d75fc3d1e58515b0c88ec6c139 Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.973049 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: W1206 15:46:00.982051 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-a497ffcfa26c7aff8bdfe411e877fdc75270398256eff6d53a3f5670bf0601da WatchSource:0}: Error finding container a497ffcfa26c7aff8bdfe411e877fdc75270398256eff6d53a3f5670bf0601da: Status 404 returned error can't find the container with id a497ffcfa26c7aff8bdfe411e877fdc75270398256eff6d53a3f5670bf0601da Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.986995 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 15:46:00 crc kubenswrapper[4813]: I1206 15:46:00.993336 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 15:46:01 crc kubenswrapper[4813]: W1206 15:46:01.011411 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-9cbdc344300eee119756a274fde4b5be09240d2a1c12e0903ca26a21fb81ef7b WatchSource:0}: Error finding container 9cbdc344300eee119756a274fde4b5be09240d2a1c12e0903ca26a21fb81ef7b: Status 404 returned error can't find the container with id 9cbdc344300eee119756a274fde4b5be09240d2a1c12e0903ca26a21fb81ef7b Dec 06 15:46:01 crc kubenswrapper[4813]: W1206 15:46:01.015288 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-096015c508078d6059f279a9ac081d7e2b21d7db86e4d939c9229a5fd9576d1b WatchSource:0}: Error finding container 096015c508078d6059f279a9ac081d7e2b21d7db86e4d939c9229a5fd9576d1b: Status 404 returned error can't find the container with id 096015c508078d6059f279a9ac081d7e2b21d7db86e4d939c9229a5fd9576d1b Dec 06 15:46:01 crc kubenswrapper[4813]: W1206 15:46:01.024358 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-e2785ebb0b7915f3fa5170be78d683f95735ac3404c0ff52f7b5cca373ca3067 WatchSource:0}: Error finding container e2785ebb0b7915f3fa5170be78d683f95735ac3404c0ff52f7b5cca373ca3067: Status 404 returned error can't find the container with id e2785ebb0b7915f3fa5170be78d683f95735ac3404c0ff52f7b5cca373ca3067 Dec 06 15:46:01 crc kubenswrapper[4813]: E1206 15:46:01.028746 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="800ms" Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.258335 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.260419 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.260472 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.260490 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.260525 4813 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 15:46:01 crc kubenswrapper[4813]: E1206 15:46:01.260992 4813 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.9:6443: connect: connection refused" node="crc" Dec 06 15:46:01 crc kubenswrapper[4813]: W1206 15:46:01.324771 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Dec 06 15:46:01 crc kubenswrapper[4813]: E1206 15:46:01.324864 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.423410 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.426464 4813 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 12:24:06.059038298 +0000 UTC Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.492048 4813 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7" exitCode=0 Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.492116 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7"} Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.492192 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9cbdc344300eee119756a274fde4b5be09240d2a1c12e0903ca26a21fb81ef7b"} Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.492292 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.493279 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.493326 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.493338 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.493706 4813 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207" exitCode=0 Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.493766 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207"} Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.493799 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a497ffcfa26c7aff8bdfe411e877fdc75270398256eff6d53a3f5670bf0601da"} Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.493927 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.494793 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.494823 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.494832 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.495446 4813 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="1e297ada3f58d22f81f4014c0d47f04b0fe2e68c1e032e9b17e85907f27f9a6a" exitCode=0 Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.495514 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"1e297ada3f58d22f81f4014c0d47f04b0fe2e68c1e032e9b17e85907f27f9a6a"} Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.495541 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"26c1cead14106af47b3089b4c2053d618b3045d75fc3d1e58515b0c88ec6c139"} Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.495617 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.496841 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.496868 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.496877 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.501805 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.502463 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.502486 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.502495 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.503089 4813 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="e9471d3db940ee579d5b60cf1f7750c2c265a8f8e549aaa7ed60e82982cf80d9" exitCode=0 Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.503137 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"e9471d3db940ee579d5b60cf1f7750c2c265a8f8e549aaa7ed60e82982cf80d9"} Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.503159 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e2785ebb0b7915f3fa5170be78d683f95735ac3404c0ff52f7b5cca373ca3067"} Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.503269 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.504306 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.504330 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.504339 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.507105 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545"} Dec 06 15:46:01 crc kubenswrapper[4813]: I1206 15:46:01.507130 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"096015c508078d6059f279a9ac081d7e2b21d7db86e4d939c9229a5fd9576d1b"} Dec 06 15:46:01 crc kubenswrapper[4813]: W1206 15:46:01.566247 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Dec 06 15:46:01 crc kubenswrapper[4813]: E1206 15:46:01.566398 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Dec 06 15:46:01 crc kubenswrapper[4813]: W1206 15:46:01.592475 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Dec 06 15:46:01 crc kubenswrapper[4813]: E1206 15:46:01.592567 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Dec 06 15:46:01 crc kubenswrapper[4813]: E1206 15:46:01.829830 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="1.6s" Dec 06 15:46:01 crc kubenswrapper[4813]: W1206 15:46:01.852709 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Dec 06 15:46:01 crc kubenswrapper[4813]: E1206 15:46:01.852774 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.063097 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.121313 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.121366 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.121375 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.121398 4813 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 15:46:02 crc kubenswrapper[4813]: E1206 15:46:02.121847 4813 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.9:6443: connect: connection refused" node="crc" Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.337673 4813 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 06 15:46:02 crc kubenswrapper[4813]: E1206 15:46:02.338751 4813 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.423124 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.427124 4813 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 14:28:10.483717898 +0000 UTC Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.427165 4813 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 94h42m8.056554964s for next certificate rotation Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.519542 4813 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066" exitCode=0 Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.519617 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066"} Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.520504 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.522231 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"23e3d7272d34a92fa558ed95d304ebaf3f6f0e79e5d60538187542d201250d93"} Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.522247 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.522320 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.522334 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.522389 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.523900 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.523922 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.523931 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.531282 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"8326b53a1a99f7cbd09d55008e9c1b4087993a25ae4b50fafc70e4c7b99138d3"} Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.531323 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"2c2f1dd12351d9f1ce3ccc9711361cf346d9aa1c4cb5d971020e242c80e557eb"} Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.531334 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"2a77b5cb2e8c50b39c3369f156a7800603e7d8bf3d6a544ad0f3a6814361338f"} Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.531448 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.532354 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.532380 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.532388 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.534026 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e"} Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.534083 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64"} Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.534094 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119"} Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.534061 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.534960 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.534979 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.534989 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.537599 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0"} Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.537640 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb"} Dec 06 15:46:02 crc kubenswrapper[4813]: I1206 15:46:02.537657 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137"} Dec 06 15:46:03 crc kubenswrapper[4813]: W1206 15:46:03.276348 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Dec 06 15:46:03 crc kubenswrapper[4813]: E1206 15:46:03.276424 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Dec 06 15:46:03 crc kubenswrapper[4813]: I1206 15:46:03.422898 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Dec 06 15:46:03 crc kubenswrapper[4813]: E1206 15:46:03.430988 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="3.2s" Dec 06 15:46:03 crc kubenswrapper[4813]: I1206 15:46:03.544986 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935"} Dec 06 15:46:03 crc kubenswrapper[4813]: I1206 15:46:03.545038 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789"} Dec 06 15:46:03 crc kubenswrapper[4813]: I1206 15:46:03.545143 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:03 crc kubenswrapper[4813]: I1206 15:46:03.546378 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:03 crc kubenswrapper[4813]: I1206 15:46:03.546408 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:03 crc kubenswrapper[4813]: I1206 15:46:03.546418 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:03 crc kubenswrapper[4813]: I1206 15:46:03.547468 4813 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76" exitCode=0 Dec 06 15:46:03 crc kubenswrapper[4813]: I1206 15:46:03.547533 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76"} Dec 06 15:46:03 crc kubenswrapper[4813]: I1206 15:46:03.547539 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 15:46:03 crc kubenswrapper[4813]: I1206 15:46:03.547637 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:03 crc kubenswrapper[4813]: I1206 15:46:03.547656 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:03 crc kubenswrapper[4813]: I1206 15:46:03.547686 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:03 crc kubenswrapper[4813]: I1206 15:46:03.548659 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:03 crc kubenswrapper[4813]: I1206 15:46:03.548672 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:03 crc kubenswrapper[4813]: I1206 15:46:03.548685 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:03 crc kubenswrapper[4813]: I1206 15:46:03.548688 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:03 crc kubenswrapper[4813]: I1206 15:46:03.548696 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:03 crc kubenswrapper[4813]: I1206 15:46:03.548697 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:03 crc kubenswrapper[4813]: I1206 15:46:03.548797 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:03 crc kubenswrapper[4813]: I1206 15:46:03.548825 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:03 crc kubenswrapper[4813]: I1206 15:46:03.548836 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:03 crc kubenswrapper[4813]: I1206 15:46:03.723602 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:03 crc kubenswrapper[4813]: I1206 15:46:03.733949 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:03 crc kubenswrapper[4813]: I1206 15:46:03.733976 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:03 crc kubenswrapper[4813]: I1206 15:46:03.733985 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:03 crc kubenswrapper[4813]: I1206 15:46:03.734009 4813 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 15:46:03 crc kubenswrapper[4813]: E1206 15:46:03.734412 4813 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.9:6443: connect: connection refused" node="crc" Dec 06 15:46:03 crc kubenswrapper[4813]: W1206 15:46:03.750796 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Dec 06 15:46:03 crc kubenswrapper[4813]: E1206 15:46:03.750882 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Dec 06 15:46:04 crc kubenswrapper[4813]: I1206 15:46:04.001075 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:46:04 crc kubenswrapper[4813]: I1206 15:46:04.423813 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Dec 06 15:46:04 crc kubenswrapper[4813]: W1206 15:46:04.530856 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Dec 06 15:46:04 crc kubenswrapper[4813]: E1206 15:46:04.530933 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Dec 06 15:46:04 crc kubenswrapper[4813]: I1206 15:46:04.553568 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:04 crc kubenswrapper[4813]: I1206 15:46:04.554242 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4"} Dec 06 15:46:04 crc kubenswrapper[4813]: I1206 15:46:04.554320 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1"} Dec 06 15:46:04 crc kubenswrapper[4813]: I1206 15:46:04.554340 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9"} Dec 06 15:46:04 crc kubenswrapper[4813]: I1206 15:46:04.554355 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9"} Dec 06 15:46:04 crc kubenswrapper[4813]: I1206 15:46:04.554388 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:46:04 crc kubenswrapper[4813]: I1206 15:46:04.581657 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:04 crc kubenswrapper[4813]: I1206 15:46:04.581711 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:04 crc kubenswrapper[4813]: I1206 15:46:04.581723 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:04 crc kubenswrapper[4813]: W1206 15:46:04.630241 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Dec 06 15:46:04 crc kubenswrapper[4813]: E1206 15:46:04.630407 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Dec 06 15:46:05 crc kubenswrapper[4813]: I1206 15:46:05.691370 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206"} Dec 06 15:46:05 crc kubenswrapper[4813]: I1206 15:46:05.691446 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:05 crc kubenswrapper[4813]: I1206 15:46:05.691472 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:05 crc kubenswrapper[4813]: I1206 15:46:05.694559 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:05 crc kubenswrapper[4813]: I1206 15:46:05.694601 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:05 crc kubenswrapper[4813]: I1206 15:46:05.694614 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:05 crc kubenswrapper[4813]: I1206 15:46:05.695121 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:05 crc kubenswrapper[4813]: I1206 15:46:05.695192 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:05 crc kubenswrapper[4813]: I1206 15:46:05.695211 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:06 crc kubenswrapper[4813]: I1206 15:46:06.695078 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:06 crc kubenswrapper[4813]: I1206 15:46:06.696674 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:06 crc kubenswrapper[4813]: I1206 15:46:06.696734 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:06 crc kubenswrapper[4813]: I1206 15:46:06.696754 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:06 crc kubenswrapper[4813]: I1206 15:46:06.725681 4813 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 06 15:46:06 crc kubenswrapper[4813]: I1206 15:46:06.847044 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:46:06 crc kubenswrapper[4813]: I1206 15:46:06.847311 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:06 crc kubenswrapper[4813]: I1206 15:46:06.848912 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:06 crc kubenswrapper[4813]: I1206 15:46:06.848978 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:06 crc kubenswrapper[4813]: I1206 15:46:06.848997 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:06 crc kubenswrapper[4813]: I1206 15:46:06.934614 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:06 crc kubenswrapper[4813]: I1206 15:46:06.936333 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:06 crc kubenswrapper[4813]: I1206 15:46:06.936394 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:06 crc kubenswrapper[4813]: I1206 15:46:06.936412 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:06 crc kubenswrapper[4813]: I1206 15:46:06.936461 4813 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 15:46:07 crc kubenswrapper[4813]: I1206 15:46:07.422624 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 15:46:07 crc kubenswrapper[4813]: I1206 15:46:07.423109 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:07 crc kubenswrapper[4813]: I1206 15:46:07.424743 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:07 crc kubenswrapper[4813]: I1206 15:46:07.425074 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:07 crc kubenswrapper[4813]: I1206 15:46:07.425246 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:10 crc kubenswrapper[4813]: I1206 15:46:10.423074 4813 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded" start-of-body= Dec 06 15:46:10 crc kubenswrapper[4813]: I1206 15:46:10.423195 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded" Dec 06 15:46:10 crc kubenswrapper[4813]: I1206 15:46:10.494510 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 06 15:46:10 crc kubenswrapper[4813]: I1206 15:46:10.494803 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:10 crc kubenswrapper[4813]: I1206 15:46:10.496406 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:10 crc kubenswrapper[4813]: I1206 15:46:10.496476 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:10 crc kubenswrapper[4813]: I1206 15:46:10.496502 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:10 crc kubenswrapper[4813]: E1206 15:46:10.561490 4813 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 06 15:46:10 crc kubenswrapper[4813]: I1206 15:46:10.640579 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 15:46:10 crc kubenswrapper[4813]: I1206 15:46:10.641067 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:10 crc kubenswrapper[4813]: I1206 15:46:10.643183 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:10 crc kubenswrapper[4813]: I1206 15:46:10.643242 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:10 crc kubenswrapper[4813]: I1206 15:46:10.643310 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:10 crc kubenswrapper[4813]: I1206 15:46:10.739562 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 15:46:10 crc kubenswrapper[4813]: I1206 15:46:10.739975 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:10 crc kubenswrapper[4813]: I1206 15:46:10.741734 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:10 crc kubenswrapper[4813]: I1206 15:46:10.741984 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:10 crc kubenswrapper[4813]: I1206 15:46:10.742139 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:11 crc kubenswrapper[4813]: I1206 15:46:11.595588 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 15:46:11 crc kubenswrapper[4813]: I1206 15:46:11.605839 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 15:46:11 crc kubenswrapper[4813]: I1206 15:46:11.708344 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:11 crc kubenswrapper[4813]: I1206 15:46:11.708615 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 15:46:11 crc kubenswrapper[4813]: I1206 15:46:11.709777 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:11 crc kubenswrapper[4813]: I1206 15:46:11.709841 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:11 crc kubenswrapper[4813]: I1206 15:46:11.709863 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:11 crc kubenswrapper[4813]: I1206 15:46:11.717484 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 15:46:12 crc kubenswrapper[4813]: I1206 15:46:12.064536 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 06 15:46:12 crc kubenswrapper[4813]: I1206 15:46:12.064768 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:12 crc kubenswrapper[4813]: I1206 15:46:12.066381 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:12 crc kubenswrapper[4813]: I1206 15:46:12.066475 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:12 crc kubenswrapper[4813]: I1206 15:46:12.066494 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:12 crc kubenswrapper[4813]: I1206 15:46:12.710868 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:12 crc kubenswrapper[4813]: I1206 15:46:12.711950 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:12 crc kubenswrapper[4813]: I1206 15:46:12.712012 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:12 crc kubenswrapper[4813]: I1206 15:46:12.712023 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:13 crc kubenswrapper[4813]: I1206 15:46:13.713315 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:13 crc kubenswrapper[4813]: I1206 15:46:13.714296 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:13 crc kubenswrapper[4813]: I1206 15:46:13.714324 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:13 crc kubenswrapper[4813]: I1206 15:46:13.714336 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:14 crc kubenswrapper[4813]: I1206 15:46:14.306863 4813 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 06 15:46:14 crc kubenswrapper[4813]: I1206 15:46:14.306920 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 06 15:46:14 crc kubenswrapper[4813]: I1206 15:46:14.312764 4813 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 06 15:46:14 crc kubenswrapper[4813]: I1206 15:46:14.312823 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 06 15:46:16 crc kubenswrapper[4813]: I1206 15:46:16.853047 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:46:16 crc kubenswrapper[4813]: I1206 15:46:16.853715 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:16 crc kubenswrapper[4813]: I1206 15:46:16.853812 4813 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 06 15:46:16 crc kubenswrapper[4813]: I1206 15:46:16.853852 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 06 15:46:16 crc kubenswrapper[4813]: I1206 15:46:16.855122 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:16 crc kubenswrapper[4813]: I1206 15:46:16.855171 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:16 crc kubenswrapper[4813]: I1206 15:46:16.855192 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:16 crc kubenswrapper[4813]: I1206 15:46:16.859963 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:46:17 crc kubenswrapper[4813]: I1206 15:46:17.725086 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:17 crc kubenswrapper[4813]: I1206 15:46:17.725413 4813 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 06 15:46:17 crc kubenswrapper[4813]: I1206 15:46:17.725501 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 06 15:46:17 crc kubenswrapper[4813]: I1206 15:46:17.726404 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:17 crc kubenswrapper[4813]: I1206 15:46:17.726448 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:17 crc kubenswrapper[4813]: I1206 15:46:17.726458 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:19 crc kubenswrapper[4813]: E1206 15:46:19.304379 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.304696 4813 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.306566 4813 trace.go:236] Trace[37016314]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 15:46:07.563) (total time: 11743ms): Dec 06 15:46:19 crc kubenswrapper[4813]: Trace[37016314]: ---"Objects listed" error: 11743ms (15:46:19.306) Dec 06 15:46:19 crc kubenswrapper[4813]: Trace[37016314]: [11.743428447s] [11.743428447s] END Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.306599 4813 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.307060 4813 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.307528 4813 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.309208 4813 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 06 15:46:19 crc kubenswrapper[4813]: E1206 15:46:19.310837 4813 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.317885 4813 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.351945 4813 csr.go:261] certificate signing request csr-pp4bv is approved, waiting to be issued Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.365831 4813 csr.go:257] certificate signing request csr-pp4bv is issued Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.399432 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.413272 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.416794 4813 apiserver.go:52] "Watching apiserver" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.420902 4813 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.421215 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.421579 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.421697 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.421845 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.421912 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.422107 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.422259 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:19 crc kubenswrapper[4813]: E1206 15:46:19.422329 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:46:19 crc kubenswrapper[4813]: E1206 15:46:19.422162 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:46:19 crc kubenswrapper[4813]: E1206 15:46:19.422097 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.443276 4813 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.443432 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.443453 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.443578 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.443711 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.443754 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.451619 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.454943 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.455016 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516367 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516408 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516430 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516451 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516468 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516487 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516503 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516517 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516533 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516548 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516562 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516576 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516590 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516605 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516620 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516635 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516651 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516667 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516682 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516697 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516713 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516730 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516750 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516769 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516785 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516799 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516815 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516830 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516845 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516862 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516877 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516892 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516917 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516932 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516950 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516965 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516980 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.516996 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517013 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517028 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517045 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517062 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517077 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517093 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517108 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517142 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517158 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517176 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517193 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517208 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517223 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517237 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517253 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517289 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517305 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517319 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517337 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517352 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517367 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517488 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517509 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517523 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517540 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517555 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517571 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517586 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517609 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517626 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517649 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517672 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517692 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517709 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517726 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517744 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517761 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517777 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517794 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517809 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517826 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517843 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517858 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517874 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517891 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517908 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517924 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517925 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.517940 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518006 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518031 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518049 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518067 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518086 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518105 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518123 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518142 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518148 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518158 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518178 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518195 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518212 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518228 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518247 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518281 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518299 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518319 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518328 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518335 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518366 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518385 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518402 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518421 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518438 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518455 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518462 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518476 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518498 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518518 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518534 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518551 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518567 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518584 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518602 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518618 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518634 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518649 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518666 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518682 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518698 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518714 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518735 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518760 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518786 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518808 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518823 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518839 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518858 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518880 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518904 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518914 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518924 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518966 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.518992 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.519010 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.519029 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.519046 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.519064 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.519086 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.519104 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.519121 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.519139 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.519155 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.519172 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.519189 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.519207 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.519227 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.519245 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.519282 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.519303 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.519321 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.519047 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.519073 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.519206 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.519216 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.519377 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.519589 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.519604 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.519813 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.519824 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.519977 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.520040 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.520086 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.520258 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.520387 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.520394 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.520393 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.520420 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.520503 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.520565 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.520587 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.520608 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.520708 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.520718 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.520744 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.520799 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.520868 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.520945 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.520974 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.520987 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.521044 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.521117 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.521152 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.521170 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.521179 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.521230 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.521395 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.521503 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.521787 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.521816 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.521834 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.521854 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.521873 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.521891 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.521910 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.521930 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.521949 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.521967 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.521987 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522006 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522023 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522040 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522062 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522080 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522098 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522116 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522135 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522176 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522195 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522213 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522231 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522250 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522581 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522603 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522623 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522642 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522669 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522690 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522708 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522726 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522743 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522759 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522775 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522793 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522812 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522829 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522846 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522867 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522885 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522934 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522960 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.523000 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.523022 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.523043 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.523063 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.523081 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.523099 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.523118 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.523137 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.523155 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522370 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522384 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.538611 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522467 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.522584 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.523352 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.523346 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.523541 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.523756 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.523777 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.523794 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.523947 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.523980 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.523978 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.524122 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.524280 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.524304 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.524340 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.524508 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.524548 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.524654 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.524686 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.524699 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.527424 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.527394 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.527545 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.527634 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.527733 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.527809 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.527822 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.527888 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.528063 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.528147 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.528198 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.528311 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.528399 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.528482 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.528714 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.528722 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.528780 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.528730 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.528973 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.528982 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.529145 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.529186 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.529265 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.529455 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.529475 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.529590 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.529667 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.529715 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.529842 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.529855 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.530340 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.530359 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.530389 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.530436 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.530634 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.530848 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.530900 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.531081 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.531092 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.531237 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.531263 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.531436 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.531452 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.531528 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.531542 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.531671 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.531715 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.531800 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: E1206 15:46:19.531878 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:46:20.030381916 +0000 UTC m=+19.921261492 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.532063 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.532908 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.532926 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.533170 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.533363 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.533632 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.533657 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.533776 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.533832 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.534543 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.534914 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.534935 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.535064 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.535109 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.535265 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.535316 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.535529 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.537191 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.537774 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.538380 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.538586 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.538781 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.538916 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.539039 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.539404 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.539422 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.539626 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.539725 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.539768 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.539791 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.539865 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.539883 4813 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.539918 4813 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.539933 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.539993 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.540532 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.540685 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.540821 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.540977 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.541582 4813 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.542420 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.542734 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.542940 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.542991 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.543297 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.543473 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.543607 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.543917 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.544032 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.544147 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.544321 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.574671 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.578392 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.578556 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.579097 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.581583 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.581779 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.581905 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.581975 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.582165 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.594716 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 15:46:19 crc kubenswrapper[4813]: E1206 15:46:19.594798 4813 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 15:46:19 crc kubenswrapper[4813]: E1206 15:46:19.594845 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 15:46:20.094830773 +0000 UTC m=+19.985710349 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.595075 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 15:46:19 crc kubenswrapper[4813]: E1206 15:46:19.595119 4813 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 15:46:19 crc kubenswrapper[4813]: E1206 15:46:19.595143 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 15:46:20.095137081 +0000 UTC m=+19.986016657 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.595339 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.595554 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.595752 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.595818 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.595981 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.596358 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.596565 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.596771 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.597179 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.597229 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.597443 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 15:46:19 crc kubenswrapper[4813]: E1206 15:46:19.597640 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 15:46:19 crc kubenswrapper[4813]: E1206 15:46:19.597663 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 15:46:19 crc kubenswrapper[4813]: E1206 15:46:19.597677 4813 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:46:19 crc kubenswrapper[4813]: E1206 15:46:19.597734 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 15:46:20.097718067 +0000 UTC m=+19.988597753 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.597796 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.598328 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.598501 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.598686 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.598819 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.599134 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.599903 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.603355 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.607283 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.609454 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.609568 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.610081 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.640631 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.640881 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641001 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641051 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641058 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641148 4813 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641163 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641177 4813 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641188 4813 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641200 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641212 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641227 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641242 4813 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641256 4813 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641285 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641298 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641310 4813 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641321 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641332 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641343 4813 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641355 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641367 4813 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641379 4813 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641393 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641405 4813 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641416 4813 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641426 4813 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641436 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641446 4813 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641457 4813 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641467 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641479 4813 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641492 4813 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641502 4813 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641514 4813 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641525 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641537 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641550 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641563 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641574 4813 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641584 4813 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641596 4813 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641608 4813 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641619 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641630 4813 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641641 4813 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641653 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641666 4813 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641679 4813 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641691 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641703 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641713 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641726 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641737 4813 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641750 4813 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641761 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641776 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641788 4813 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641799 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641810 4813 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641821 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641833 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641844 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641856 4813 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641867 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641878 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641891 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641903 4813 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641914 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641927 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641939 4813 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641951 4813 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641964 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641977 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.641990 4813 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642002 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642016 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642027 4813 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642039 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642051 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642065 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642077 4813 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642088 4813 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642099 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642112 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642124 4813 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642134 4813 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642146 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642158 4813 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642170 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642181 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642192 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642204 4813 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.640808 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642230 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642290 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642305 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642315 4813 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642327 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642339 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642354 4813 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642368 4813 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642381 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642393 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642401 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642410 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642419 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642428 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642438 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642449 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642461 4813 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642473 4813 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642484 4813 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642493 4813 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642501 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642511 4813 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642520 4813 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642531 4813 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642543 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642599 4813 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642610 4813 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642618 4813 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642626 4813 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642636 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642645 4813 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642653 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642663 4813 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642671 4813 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642680 4813 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642689 4813 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642697 4813 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642707 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642715 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642724 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642733 4813 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642743 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642750 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642759 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642768 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642777 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642784 4813 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642792 4813 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642800 4813 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642808 4813 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642816 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642823 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642832 4813 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642841 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642848 4813 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642857 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642865 4813 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642874 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642885 4813 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642896 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642908 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642917 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642932 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642941 4813 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642950 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642958 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642968 4813 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642976 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642985 4813 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.642993 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.643001 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.643010 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.643019 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.643028 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.643036 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.643045 4813 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.643052 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.643062 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.643070 4813 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.643079 4813 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.643087 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.643095 4813 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.643102 4813 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.643112 4813 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.643120 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.643129 4813 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.643136 4813 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.665014 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.665189 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.668732 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.671479 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.671604 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.671807 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.671998 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.672135 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.673900 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.676748 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.681028 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.682775 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.690690 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.692647 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.793694 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.793898 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.795466 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.795481 4813 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.795491 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.795500 4813 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.795508 4813 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.795517 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.795526 4813 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.795535 4813 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.795543 4813 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.795554 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.793998 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.813594 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:46:19 crc kubenswrapper[4813]: W1206 15:46:19.815559 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-6c23062d894b90fea136100a6f9e044a3edc2ef237ef100e6df6d24465739c1d WatchSource:0}: Error finding container 6c23062d894b90fea136100a6f9e044a3edc2ef237ef100e6df6d24465739c1d: Status 404 returned error can't find the container with id 6c23062d894b90fea136100a6f9e044a3edc2ef237ef100e6df6d24465739c1d Dec 06 15:46:19 crc kubenswrapper[4813]: E1206 15:46:19.819608 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 15:46:19 crc kubenswrapper[4813]: E1206 15:46:19.819693 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 15:46:19 crc kubenswrapper[4813]: E1206 15:46:19.819767 4813 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:46:19 crc kubenswrapper[4813]: E1206 15:46:19.819865 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 15:46:20.319847779 +0000 UTC m=+20.210727355 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:46:19 crc kubenswrapper[4813]: W1206 15:46:19.820004 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-25ec78ee76184f66fd8fc4119c4f7c46b2570de461b8adcbf65382282930d8ff WatchSource:0}: Error finding container 25ec78ee76184f66fd8fc4119c4f7c46b2570de461b8adcbf65382282930d8ff: Status 404 returned error can't find the container with id 25ec78ee76184f66fd8fc4119c4f7c46b2570de461b8adcbf65382282930d8ff Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.855867 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.856822 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.895813 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.900766 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.915574 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.927859 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.944315 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 15:46:19 crc kubenswrapper[4813]: I1206 15:46:19.973907 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.036047 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.056433 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.097624 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.097699 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.097731 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.097755 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:20 crc kubenswrapper[4813]: E1206 15:46:20.097840 4813 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 15:46:20 crc kubenswrapper[4813]: E1206 15:46:20.097882 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 15:46:21.097869118 +0000 UTC m=+20.988748694 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 15:46:20 crc kubenswrapper[4813]: E1206 15:46:20.097929 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:46:21.09792427 +0000 UTC m=+20.988803846 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:46:20 crc kubenswrapper[4813]: E1206 15:46:20.097954 4813 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 15:46:20 crc kubenswrapper[4813]: E1206 15:46:20.097972 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 15:46:21.097967391 +0000 UTC m=+20.988846967 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 15:46:20 crc kubenswrapper[4813]: E1206 15:46:20.098022 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 15:46:20 crc kubenswrapper[4813]: E1206 15:46:20.098034 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 15:46:20 crc kubenswrapper[4813]: E1206 15:46:20.098044 4813 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:46:20 crc kubenswrapper[4813]: E1206 15:46:20.098070 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 15:46:21.098062513 +0000 UTC m=+20.988942089 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.222485 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-t5xp8"] Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.222860 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.223681 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-t9zmh"] Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.224142 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-mhncd"] Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.224213 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.224379 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-mhncd" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.224960 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-gl495"] Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.225157 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.226168 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.226707 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.226893 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.227132 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.227294 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.227440 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.227552 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.227648 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.227733 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.227826 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.230465 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.231230 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.231358 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.231466 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.231653 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.238609 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.255784 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.280328 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.296110 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.299333 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/6adb3a02-198b-41b4-8b18-56b0f9527fe7-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-t9zmh\" (UID: \"6adb3a02-198b-41b4-8b18-56b0f9527fe7\") " pod="openshift-multus/multus-additional-cni-plugins-t9zmh" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.299369 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-host-run-k8s-cni-cncf-io\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.299399 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6adb3a02-198b-41b4-8b18-56b0f9527fe7-cni-binary-copy\") pod \"multus-additional-cni-plugins-t9zmh\" (UID: \"6adb3a02-198b-41b4-8b18-56b0f9527fe7\") " pod="openshift-multus/multus-additional-cni-plugins-t9zmh" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.299424 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6ddx\" (UniqueName: \"kubernetes.io/projected/8dd11f51-430a-4e7a-ab55-9666abfb1e82-kube-api-access-f6ddx\") pod \"node-resolver-mhncd\" (UID: \"8dd11f51-430a-4e7a-ab55-9666abfb1e82\") " pod="openshift-dns/node-resolver-mhncd" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.299442 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-os-release\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.299458 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-multus-socket-dir-parent\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.299476 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-host-var-lib-kubelet\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.299520 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6adb3a02-198b-41b4-8b18-56b0f9527fe7-tuning-conf-dir\") pod \"multus-additional-cni-plugins-t9zmh\" (UID: \"6adb3a02-198b-41b4-8b18-56b0f9527fe7\") " pod="openshift-multus/multus-additional-cni-plugins-t9zmh" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.299535 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-multus-cni-dir\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.299553 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/856bb45b-525c-4ef0-bf7c-0691cf54b342-cni-binary-copy\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.299585 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-multus-conf-dir\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.299608 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d88e8bae-c055-4c55-b548-f621ff96de06-proxy-tls\") pod \"machine-config-daemon-t5xp8\" (UID: \"d88e8bae-c055-4c55-b548-f621ff96de06\") " pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.299632 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5246p\" (UniqueName: \"kubernetes.io/projected/856bb45b-525c-4ef0-bf7c-0691cf54b342-kube-api-access-5246p\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.299647 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cg64b\" (UniqueName: \"kubernetes.io/projected/d88e8bae-c055-4c55-b548-f621ff96de06-kube-api-access-cg64b\") pod \"machine-config-daemon-t5xp8\" (UID: \"d88e8bae-c055-4c55-b548-f621ff96de06\") " pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.299665 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6adb3a02-198b-41b4-8b18-56b0f9527fe7-system-cni-dir\") pod \"multus-additional-cni-plugins-t9zmh\" (UID: \"6adb3a02-198b-41b4-8b18-56b0f9527fe7\") " pod="openshift-multus/multus-additional-cni-plugins-t9zmh" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.299686 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-hostroot\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.299703 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-cnibin\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.299716 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-host-var-lib-cni-bin\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.299736 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/856bb45b-525c-4ef0-bf7c-0691cf54b342-multus-daemon-config\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.299749 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d88e8bae-c055-4c55-b548-f621ff96de06-rootfs\") pod \"machine-config-daemon-t5xp8\" (UID: \"d88e8bae-c055-4c55-b548-f621ff96de06\") " pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.299764 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6adb3a02-198b-41b4-8b18-56b0f9527fe7-cnibin\") pod \"multus-additional-cni-plugins-t9zmh\" (UID: \"6adb3a02-198b-41b4-8b18-56b0f9527fe7\") " pod="openshift-multus/multus-additional-cni-plugins-t9zmh" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.299779 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-host-run-netns\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.299798 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-system-cni-dir\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.299811 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6adb3a02-198b-41b4-8b18-56b0f9527fe7-os-release\") pod \"multus-additional-cni-plugins-t9zmh\" (UID: \"6adb3a02-198b-41b4-8b18-56b0f9527fe7\") " pod="openshift-multus/multus-additional-cni-plugins-t9zmh" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.299827 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k49xc\" (UniqueName: \"kubernetes.io/projected/6adb3a02-198b-41b4-8b18-56b0f9527fe7-kube-api-access-k49xc\") pod \"multus-additional-cni-plugins-t9zmh\" (UID: \"6adb3a02-198b-41b4-8b18-56b0f9527fe7\") " pod="openshift-multus/multus-additional-cni-plugins-t9zmh" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.299841 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-host-var-lib-cni-multus\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.299855 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-host-run-multus-certs\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.299868 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-etc-kubernetes\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.299883 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8dd11f51-430a-4e7a-ab55-9666abfb1e82-hosts-file\") pod \"node-resolver-mhncd\" (UID: \"8dd11f51-430a-4e7a-ab55-9666abfb1e82\") " pod="openshift-dns/node-resolver-mhncd" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.299896 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d88e8bae-c055-4c55-b548-f621ff96de06-mcd-auth-proxy-config\") pod \"machine-config-daemon-t5xp8\" (UID: \"d88e8bae-c055-4c55-b548-f621ff96de06\") " pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.315223 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.326828 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.328153 4813 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 06 15:46:20 crc kubenswrapper[4813]: W1206 15:46:20.328335 4813 reflector.go:484] object-"openshift-network-operator"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 06 15:46:20 crc kubenswrapper[4813]: W1206 15:46:20.328359 4813 reflector.go:484] object-"openshift-multus"/"default-dockercfg-2q5b6": watch of *v1.Secret ended with: very short watch: object-"openshift-multus"/"default-dockercfg-2q5b6": Unexpected watch close - watch lasted less than a second and no items received Dec 06 15:46:20 crc kubenswrapper[4813]: W1206 15:46:20.328393 4813 reflector.go:484] object-"openshift-network-operator"/"iptables-alerter-script": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"iptables-alerter-script": Unexpected watch close - watch lasted less than a second and no items received Dec 06 15:46:20 crc kubenswrapper[4813]: W1206 15:46:20.328417 4813 reflector.go:484] object-"openshift-multus"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 06 15:46:20 crc kubenswrapper[4813]: W1206 15:46:20.328685 4813 reflector.go:484] object-"openshift-network-node-identity"/"env-overrides": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"env-overrides": Unexpected watch close - watch lasted less than a second and no items received Dec 06 15:46:20 crc kubenswrapper[4813]: W1206 15:46:20.328723 4813 reflector.go:484] object-"openshift-network-node-identity"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 06 15:46:20 crc kubenswrapper[4813]: W1206 15:46:20.328749 4813 reflector.go:484] object-"openshift-multus"/"multus-daemon-config": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"multus-daemon-config": Unexpected watch close - watch lasted less than a second and no items received Dec 06 15:46:20 crc kubenswrapper[4813]: W1206 15:46:20.328772 4813 reflector.go:484] object-"openshift-network-node-identity"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 06 15:46:20 crc kubenswrapper[4813]: W1206 15:46:20.328795 4813 reflector.go:484] object-"openshift-network-operator"/"metrics-tls": watch of *v1.Secret ended with: very short watch: object-"openshift-network-operator"/"metrics-tls": Unexpected watch close - watch lasted less than a second and no items received Dec 06 15:46:20 crc kubenswrapper[4813]: W1206 15:46:20.328814 4813 reflector.go:484] object-"openshift-machine-config-operator"/"proxy-tls": watch of *v1.Secret ended with: very short watch: object-"openshift-machine-config-operator"/"proxy-tls": Unexpected watch close - watch lasted less than a second and no items received Dec 06 15:46:20 crc kubenswrapper[4813]: W1206 15:46:20.328834 4813 reflector.go:484] object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": watch of *v1.Secret ended with: very short watch: object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": Unexpected watch close - watch lasted less than a second and no items received Dec 06 15:46:20 crc kubenswrapper[4813]: W1206 15:46:20.328854 4813 reflector.go:484] object-"openshift-dns"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 06 15:46:20 crc kubenswrapper[4813]: W1206 15:46:20.328874 4813 reflector.go:484] object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": watch of *v1.Secret ended with: very short watch: object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": Unexpected watch close - watch lasted less than a second and no items received Dec 06 15:46:20 crc kubenswrapper[4813]: W1206 15:46:20.328891 4813 reflector.go:484] object-"openshift-network-node-identity"/"ovnkube-identity-cm": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"ovnkube-identity-cm": Unexpected watch close - watch lasted less than a second and no items received Dec 06 15:46:20 crc kubenswrapper[4813]: W1206 15:46:20.328930 4813 reflector.go:484] object-"openshift-network-operator"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 06 15:46:20 crc kubenswrapper[4813]: W1206 15:46:20.328948 4813 reflector.go:484] object-"openshift-machine-config-operator"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 06 15:46:20 crc kubenswrapper[4813]: W1206 15:46:20.328966 4813 reflector.go:484] object-"openshift-machine-config-operator"/"kube-rbac-proxy": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"kube-rbac-proxy": Unexpected watch close - watch lasted less than a second and no items received Dec 06 15:46:20 crc kubenswrapper[4813]: W1206 15:46:20.328987 4813 reflector.go:484] object-"openshift-machine-config-operator"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 06 15:46:20 crc kubenswrapper[4813]: W1206 15:46:20.328338 4813 reflector.go:484] object-"openshift-multus"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 06 15:46:20 crc kubenswrapper[4813]: W1206 15:46:20.329009 4813 reflector.go:484] object-"openshift-multus"/"default-cni-sysctl-allowlist": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"default-cni-sysctl-allowlist": Unexpected watch close - watch lasted less than a second and no items received Dec 06 15:46:20 crc kubenswrapper[4813]: W1206 15:46:20.329030 4813 reflector.go:484] object-"openshift-network-node-identity"/"network-node-identity-cert": watch of *v1.Secret ended with: very short watch: object-"openshift-network-node-identity"/"network-node-identity-cert": Unexpected watch close - watch lasted less than a second and no items received Dec 06 15:46:20 crc kubenswrapper[4813]: W1206 15:46:20.329049 4813 reflector.go:484] object-"openshift-multus"/"cni-copy-resources": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"cni-copy-resources": Unexpected watch close - watch lasted less than a second and no items received Dec 06 15:46:20 crc kubenswrapper[4813]: W1206 15:46:20.329066 4813 reflector.go:484] object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": watch of *v1.Secret ended with: very short watch: object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": Unexpected watch close - watch lasted less than a second and no items received Dec 06 15:46:20 crc kubenswrapper[4813]: E1206 15:46:20.328686 4813 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/events\": read tcp 38.102.83.9:55706->38.102.83.9:6443: use of closed network connection" event=< Dec 06 15:46:20 crc kubenswrapper[4813]: &Event{ObjectMeta:{kube-controller-manager-crc.187eaadd10dbf097 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": context deadline exceeded Dec 06 15:46:20 crc kubenswrapper[4813]: body: Dec 06 15:46:20 crc kubenswrapper[4813]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 15:46:10.423156887 +0000 UTC m=+10.314036503,LastTimestamp:2025-12-06 15:46:10.423156887 +0000 UTC m=+10.314036503,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Dec 06 15:46:20 crc kubenswrapper[4813]: > Dec 06 15:46:20 crc kubenswrapper[4813]: W1206 15:46:20.329092 4813 reflector.go:484] object-"openshift-dns"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.368754 4813 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-06 15:41:19 +0000 UTC, rotation deadline is 2026-10-15 08:30:07.062243883 +0000 UTC Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.368831 4813 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7504h43m46.693416657s for next certificate rotation Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.401169 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6adb3a02-198b-41b4-8b18-56b0f9527fe7-tuning-conf-dir\") pod \"multus-additional-cni-plugins-t9zmh\" (UID: \"6adb3a02-198b-41b4-8b18-56b0f9527fe7\") " pod="openshift-multus/multus-additional-cni-plugins-t9zmh" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.401332 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-multus-cni-dir\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.401356 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/856bb45b-525c-4ef0-bf7c-0691cf54b342-cni-binary-copy\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.401370 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-multus-conf-dir\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.401448 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d88e8bae-c055-4c55-b548-f621ff96de06-proxy-tls\") pod \"machine-config-daemon-t5xp8\" (UID: \"d88e8bae-c055-4c55-b548-f621ff96de06\") " pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.401523 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cg64b\" (UniqueName: \"kubernetes.io/projected/d88e8bae-c055-4c55-b548-f621ff96de06-kube-api-access-cg64b\") pod \"machine-config-daemon-t5xp8\" (UID: \"d88e8bae-c055-4c55-b548-f621ff96de06\") " pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.401551 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6adb3a02-198b-41b4-8b18-56b0f9527fe7-system-cni-dir\") pod \"multus-additional-cni-plugins-t9zmh\" (UID: \"6adb3a02-198b-41b4-8b18-56b0f9527fe7\") " pod="openshift-multus/multus-additional-cni-plugins-t9zmh" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.401574 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.401593 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-hostroot\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.401732 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5246p\" (UniqueName: \"kubernetes.io/projected/856bb45b-525c-4ef0-bf7c-0691cf54b342-kube-api-access-5246p\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.401752 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-host-var-lib-cni-bin\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.401888 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-cnibin\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.401912 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d88e8bae-c055-4c55-b548-f621ff96de06-rootfs\") pod \"machine-config-daemon-t5xp8\" (UID: \"d88e8bae-c055-4c55-b548-f621ff96de06\") " pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.402041 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6adb3a02-198b-41b4-8b18-56b0f9527fe7-cnibin\") pod \"multus-additional-cni-plugins-t9zmh\" (UID: \"6adb3a02-198b-41b4-8b18-56b0f9527fe7\") " pod="openshift-multus/multus-additional-cni-plugins-t9zmh" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.402064 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-host-run-netns\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.402082 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/856bb45b-525c-4ef0-bf7c-0691cf54b342-multus-daemon-config\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.402204 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-system-cni-dir\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.402221 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6adb3a02-198b-41b4-8b18-56b0f9527fe7-os-release\") pod \"multus-additional-cni-plugins-t9zmh\" (UID: \"6adb3a02-198b-41b4-8b18-56b0f9527fe7\") " pod="openshift-multus/multus-additional-cni-plugins-t9zmh" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.402239 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-host-var-lib-cni-multus\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.402320 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-host-run-multus-certs\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.402338 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-etc-kubernetes\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.402384 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k49xc\" (UniqueName: \"kubernetes.io/projected/6adb3a02-198b-41b4-8b18-56b0f9527fe7-kube-api-access-k49xc\") pod \"multus-additional-cni-plugins-t9zmh\" (UID: \"6adb3a02-198b-41b4-8b18-56b0f9527fe7\") " pod="openshift-multus/multus-additional-cni-plugins-t9zmh" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.402402 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8dd11f51-430a-4e7a-ab55-9666abfb1e82-hosts-file\") pod \"node-resolver-mhncd\" (UID: \"8dd11f51-430a-4e7a-ab55-9666abfb1e82\") " pod="openshift-dns/node-resolver-mhncd" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.402421 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d88e8bae-c055-4c55-b548-f621ff96de06-mcd-auth-proxy-config\") pod \"machine-config-daemon-t5xp8\" (UID: \"d88e8bae-c055-4c55-b548-f621ff96de06\") " pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.402503 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-host-run-k8s-cni-cncf-io\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.402520 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6adb3a02-198b-41b4-8b18-56b0f9527fe7-cni-binary-copy\") pod \"multus-additional-cni-plugins-t9zmh\" (UID: \"6adb3a02-198b-41b4-8b18-56b0f9527fe7\") " pod="openshift-multus/multus-additional-cni-plugins-t9zmh" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.402622 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/6adb3a02-198b-41b4-8b18-56b0f9527fe7-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-t9zmh\" (UID: \"6adb3a02-198b-41b4-8b18-56b0f9527fe7\") " pod="openshift-multus/multus-additional-cni-plugins-t9zmh" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.402647 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-os-release\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.402733 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-multus-socket-dir-parent\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.402751 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-host-var-lib-kubelet\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.402819 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6ddx\" (UniqueName: \"kubernetes.io/projected/8dd11f51-430a-4e7a-ab55-9666abfb1e82-kube-api-access-f6ddx\") pod \"node-resolver-mhncd\" (UID: \"8dd11f51-430a-4e7a-ab55-9666abfb1e82\") " pod="openshift-dns/node-resolver-mhncd" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.409354 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-host-run-netns\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.409412 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8dd11f51-430a-4e7a-ab55-9666abfb1e82-hosts-file\") pod \"node-resolver-mhncd\" (UID: \"8dd11f51-430a-4e7a-ab55-9666abfb1e82\") " pod="openshift-dns/node-resolver-mhncd" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.409555 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-system-cni-dir\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.409625 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6adb3a02-198b-41b4-8b18-56b0f9527fe7-os-release\") pod \"multus-additional-cni-plugins-t9zmh\" (UID: \"6adb3a02-198b-41b4-8b18-56b0f9527fe7\") " pod="openshift-multus/multus-additional-cni-plugins-t9zmh" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.409665 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6adb3a02-198b-41b4-8b18-56b0f9527fe7-tuning-conf-dir\") pod \"multus-additional-cni-plugins-t9zmh\" (UID: \"6adb3a02-198b-41b4-8b18-56b0f9527fe7\") " pod="openshift-multus/multus-additional-cni-plugins-t9zmh" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.409679 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-host-var-lib-cni-multus\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.409714 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-host-run-multus-certs\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.409751 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-multus-cni-dir\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.409765 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-etc-kubernetes\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.409940 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/6adb3a02-198b-41b4-8b18-56b0f9527fe7-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-t9zmh\" (UID: \"6adb3a02-198b-41b4-8b18-56b0f9527fe7\") " pod="openshift-multus/multus-additional-cni-plugins-t9zmh" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.410171 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-multus-socket-dir-parent\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.410213 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/856bb45b-525c-4ef0-bf7c-0691cf54b342-cni-binary-copy\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.410249 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-os-release\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.410287 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-multus-conf-dir\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.410293 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/856bb45b-525c-4ef0-bf7c-0691cf54b342-multus-daemon-config\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.410326 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-host-var-lib-kubelet\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.410377 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-host-run-k8s-cni-cncf-io\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.410473 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d88e8bae-c055-4c55-b548-f621ff96de06-mcd-auth-proxy-config\") pod \"machine-config-daemon-t5xp8\" (UID: \"d88e8bae-c055-4c55-b548-f621ff96de06\") " pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.410544 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6adb3a02-198b-41b4-8b18-56b0f9527fe7-system-cni-dir\") pod \"multus-additional-cni-plugins-t9zmh\" (UID: \"6adb3a02-198b-41b4-8b18-56b0f9527fe7\") " pod="openshift-multus/multus-additional-cni-plugins-t9zmh" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.410636 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-cnibin\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: E1206 15:46:20.410663 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.410682 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d88e8bae-c055-4c55-b548-f621ff96de06-rootfs\") pod \"machine-config-daemon-t5xp8\" (UID: \"d88e8bae-c055-4c55-b548-f621ff96de06\") " pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" Dec 06 15:46:20 crc kubenswrapper[4813]: E1206 15:46:20.410693 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.410715 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-host-var-lib-cni-bin\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: E1206 15:46:20.410730 4813 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:46:20 crc kubenswrapper[4813]: E1206 15:46:20.410786 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 15:46:21.410764224 +0000 UTC m=+21.301643800 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.410780 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/856bb45b-525c-4ef0-bf7c-0691cf54b342-hostroot\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.410832 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6adb3a02-198b-41b4-8b18-56b0f9527fe7-cnibin\") pod \"multus-additional-cni-plugins-t9zmh\" (UID: \"6adb3a02-198b-41b4-8b18-56b0f9527fe7\") " pod="openshift-multus/multus-additional-cni-plugins-t9zmh" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.410885 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6adb3a02-198b-41b4-8b18-56b0f9527fe7-cni-binary-copy\") pod \"multus-additional-cni-plugins-t9zmh\" (UID: \"6adb3a02-198b-41b4-8b18-56b0f9527fe7\") " pod="openshift-multus/multus-additional-cni-plugins-t9zmh" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.415114 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d88e8bae-c055-4c55-b548-f621ff96de06-proxy-tls\") pod \"machine-config-daemon-t5xp8\" (UID: \"d88e8bae-c055-4c55-b548-f621ff96de06\") " pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.431878 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6ddx\" (UniqueName: \"kubernetes.io/projected/8dd11f51-430a-4e7a-ab55-9666abfb1e82-kube-api-access-f6ddx\") pod \"node-resolver-mhncd\" (UID: \"8dd11f51-430a-4e7a-ab55-9666abfb1e82\") " pod="openshift-dns/node-resolver-mhncd" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.433284 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cg64b\" (UniqueName: \"kubernetes.io/projected/d88e8bae-c055-4c55-b548-f621ff96de06-kube-api-access-cg64b\") pod \"machine-config-daemon-t5xp8\" (UID: \"d88e8bae-c055-4c55-b548-f621ff96de06\") " pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.433713 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k49xc\" (UniqueName: \"kubernetes.io/projected/6adb3a02-198b-41b4-8b18-56b0f9527fe7-kube-api-access-k49xc\") pod \"multus-additional-cni-plugins-t9zmh\" (UID: \"6adb3a02-198b-41b4-8b18-56b0f9527fe7\") " pod="openshift-multus/multus-additional-cni-plugins-t9zmh" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.434515 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5246p\" (UniqueName: \"kubernetes.io/projected/856bb45b-525c-4ef0-bf7c-0691cf54b342-kube-api-access-5246p\") pod \"multus-gl495\" (UID: \"856bb45b-525c-4ef0-bf7c-0691cf54b342\") " pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.491141 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.491860 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.492744 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.493622 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.494216 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.495595 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.496376 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.497618 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.498441 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.499616 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.500204 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.501595 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.502241 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.502967 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.503928 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.504509 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.505587 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.506061 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.506751 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.507772 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.508512 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.509577 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.510044 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.511140 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.511708 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.512423 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.513678 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.514475 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.515589 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.516091 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.517091 4813 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.517303 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.518961 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.519922 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.520479 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.522071 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.522954 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.524198 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.524966 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.526175 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.526785 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.527952 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.528758 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.529907 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.530592 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.531650 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.532206 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.533360 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.533915 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.534838 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.535357 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.536339 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.536957 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.537601 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.541596 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" Dec 06 15:46:20 crc kubenswrapper[4813]: W1206 15:46:20.551801 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd88e8bae_c055_4c55_b548_f621ff96de06.slice/crio-5bb2250bfa4d0d4224eb8b961d768747557c85049420704d37f04ad79cc8321b WatchSource:0}: Error finding container 5bb2250bfa4d0d4224eb8b961d768747557c85049420704d37f04ad79cc8321b: Status 404 returned error can't find the container with id 5bb2250bfa4d0d4224eb8b961d768747557c85049420704d37f04ad79cc8321b Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.562760 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.573690 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-mhncd" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.586467 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-gl495" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.843861 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.850870 4813 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935" exitCode=255 Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.850931 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935"} Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.852557 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gl495" event={"ID":"856bb45b-525c-4ef0-bf7c-0691cf54b342","Type":"ContainerStarted","Data":"fe2ea95dc5cfa3d7f1bec679b4e6faed0ec6a2f126a6174c2d0239203c93e93c"} Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.856285 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a"} Dec 06 15:46:20 crc kubenswrapper[4813]: I1206 15:46:20.856313 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"c05d7cbeab56f523a7ea816c6740188904fb4891d295a6905c8e36e07725709e"} Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.038913 4813 scope.go:117] "RemoveContainer" containerID="20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.047651 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.048569 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b"} Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.048605 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5"} Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.048615 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"6c23062d894b90fea136100a6f9e044a3edc2ef237ef100e6df6d24465739c1d"} Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.051133 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-mhncd" event={"ID":"8dd11f51-430a-4e7a-ab55-9666abfb1e82","Type":"ContainerStarted","Data":"dcb1bec6f5e8c2c0a485046bededc4aca9c04e029b2bc03502e8bcfd4fe4d3f9"} Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.076160 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" event={"ID":"6adb3a02-198b-41b4-8b18-56b0f9527fe7","Type":"ContainerStarted","Data":"d5ccd4903200bc611e8b87e19b35b09ce8dcbae2db030192ada524285f266ca2"} Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.105087 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerStarted","Data":"5bb2250bfa4d0d4224eb8b961d768747557c85049420704d37f04ad79cc8321b"} Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.121468 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"25ec78ee76184f66fd8fc4119c4f7c46b2570de461b8adcbf65382282930d8ff"} Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.141567 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:46:21 crc kubenswrapper[4813]: E1206 15:46:21.141673 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:46:23.141655868 +0000 UTC m=+23.032535444 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.141783 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:21 crc kubenswrapper[4813]: E1206 15:46:21.141911 4813 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 15:46:21 crc kubenswrapper[4813]: E1206 15:46:21.141953 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 15:46:23.141945836 +0000 UTC m=+23.032825412 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.141818 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:21 crc kubenswrapper[4813]: E1206 15:46:21.142302 4813 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 15:46:21 crc kubenswrapper[4813]: E1206 15:46:21.142325 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 15:46:23.142318245 +0000 UTC m=+23.033197821 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.142370 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:46:21 crc kubenswrapper[4813]: E1206 15:46:21.142730 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 15:46:21 crc kubenswrapper[4813]: E1206 15:46:21.142752 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 15:46:21 crc kubenswrapper[4813]: E1206 15:46:21.142763 4813 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:46:21 crc kubenswrapper[4813]: E1206 15:46:21.142786 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 15:46:23.142778737 +0000 UTC m=+23.033658313 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.147826 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.160433 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.178966 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.203634 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.223183 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.228583 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.244067 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.265083 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.393878 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.408814 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-xdt4d"] Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.409574 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.415493 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.415682 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.415775 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.415911 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:21Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.416010 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.419389 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.422362 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.432407 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.433946 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:21Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.450618 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-systemd-units\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.450662 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-cni-netd\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.450691 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0e54e31f-e9f8-4dd5-8738-c785df002c93-env-overrides\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.450713 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0e54e31f-e9f8-4dd5-8738-c785df002c93-ovn-node-metrics-cert\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.450737 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-run-ovn\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.450760 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-node-log\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.450782 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbvf9\" (UniqueName: \"kubernetes.io/projected/0e54e31f-e9f8-4dd5-8738-c785df002c93-kube-api-access-nbvf9\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.450803 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-run-netns\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.450828 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0e54e31f-e9f8-4dd5-8738-c785df002c93-ovnkube-config\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.450852 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-etc-openvswitch\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.450874 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-run-ovn-kubernetes\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.450896 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.450921 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-var-lib-openvswitch\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.450944 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-kubelet\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.450965 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-slash\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.450997 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0e54e31f-e9f8-4dd5-8738-c785df002c93-ovnkube-script-lib\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.451027 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-cni-bin\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.451053 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.451075 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-log-socket\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.451109 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-run-systemd\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.451132 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-run-openvswitch\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: E1206 15:46:21.451359 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 15:46:21 crc kubenswrapper[4813]: E1206 15:46:21.451389 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 15:46:21 crc kubenswrapper[4813]: E1206 15:46:21.451403 4813 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:46:21 crc kubenswrapper[4813]: E1206 15:46:21.451447 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 15:46:23.451432644 +0000 UTC m=+23.342312290 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.453585 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.466817 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.483573 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.485719 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.485817 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.485854 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:46:21 crc kubenswrapper[4813]: E1206 15:46:21.485872 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:46:21 crc kubenswrapper[4813]: E1206 15:46:21.485922 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:46:21 crc kubenswrapper[4813]: E1206 15:46:21.485990 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.488854 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:21Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.551877 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-kubelet\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.551918 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-var-lib-openvswitch\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.551943 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-slash\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.551968 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0e54e31f-e9f8-4dd5-8738-c785df002c93-ovnkube-script-lib\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.551994 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-cni-bin\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.552005 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-var-lib-openvswitch\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.552077 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-slash\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.552027 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-log-socket\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.552131 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-run-systemd\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.552167 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-run-openvswitch\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.552190 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-systemd-units\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.552209 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-cni-netd\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.552234 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0e54e31f-e9f8-4dd5-8738-c785df002c93-env-overrides\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.552256 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0e54e31f-e9f8-4dd5-8738-c785df002c93-ovn-node-metrics-cert\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.552295 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-run-ovn\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.552320 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-node-log\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.552341 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbvf9\" (UniqueName: \"kubernetes.io/projected/0e54e31f-e9f8-4dd5-8738-c785df002c93-kube-api-access-nbvf9\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.552361 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-run-netns\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.552389 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-etc-openvswitch\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.552409 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-run-ovn-kubernetes\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.552432 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.552454 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0e54e31f-e9f8-4dd5-8738-c785df002c93-ovnkube-config\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.552662 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0e54e31f-e9f8-4dd5-8738-c785df002c93-ovnkube-script-lib\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.552703 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-cni-bin\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.553148 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0e54e31f-e9f8-4dd5-8738-c785df002c93-ovnkube-config\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.553601 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-run-openvswitch\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.553627 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-log-socket\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.553644 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-run-systemd\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.553667 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-run-netns\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.553681 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-run-ovn\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.553696 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-node-log\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.553874 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-cni-netd\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.553899 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-systemd-units\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.554254 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0e54e31f-e9f8-4dd5-8738-c785df002c93-env-overrides\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.554333 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-run-ovn-kubernetes\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.554355 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-etc-openvswitch\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.554379 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.554400 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-kubelet\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.557199 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0e54e31f-e9f8-4dd5-8738-c785df002c93-ovn-node-metrics-cert\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.597093 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.667884 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.667886 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.669570 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.670360 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.670538 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.670643 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.678346 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:21Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.685097 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbvf9\" (UniqueName: \"kubernetes.io/projected/0e54e31f-e9f8-4dd5-8738-c785df002c93-kube-api-access-nbvf9\") pod \"ovnkube-node-xdt4d\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.695144 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:21Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.710295 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:21Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.722169 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:21Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.726119 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.742379 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:21Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.761898 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:21Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.777822 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.783042 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:21Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.802475 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:21Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.804477 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.818922 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:21Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.837635 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:21Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.853203 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:21Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.867030 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:21Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.882425 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.887228 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:21Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.901090 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:21Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.910909 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:21Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.929204 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.929753 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.934890 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.938722 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:21Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.965942 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:21Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.982383 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:21Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:21 crc kubenswrapper[4813]: I1206 15:46:21.997703 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:21Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.013102 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:22Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.024017 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:22Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.035038 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:22Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.051530 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:22Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.065818 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:22Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.079464 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:22Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.084684 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.099591 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.102030 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:22Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.105804 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.114505 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:22Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.124481 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gl495" event={"ID":"856bb45b-525c-4ef0-bf7c-0691cf54b342","Type":"ContainerStarted","Data":"e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c"} Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.126097 4813 generic.go:334] "Generic (PLEG): container finished" podID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerID="4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014" exitCode=0 Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.126145 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" event={"ID":"0e54e31f-e9f8-4dd5-8738-c785df002c93","Type":"ContainerDied","Data":"4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014"} Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.126162 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" event={"ID":"0e54e31f-e9f8-4dd5-8738-c785df002c93","Type":"ContainerStarted","Data":"4ab555fa8c335cb15fd9727b52dd5a7cb19d2550221f6cd57f88520d7f2e72a8"} Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.128295 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.133909 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1"} Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.134202 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.135317 4813 generic.go:334] "Generic (PLEG): container finished" podID="6adb3a02-198b-41b4-8b18-56b0f9527fe7" containerID="edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85" exitCode=0 Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.135409 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" event={"ID":"6adb3a02-198b-41b4-8b18-56b0f9527fe7","Type":"ContainerDied","Data":"edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85"} Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.139336 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerStarted","Data":"281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c"} Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.139372 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerStarted","Data":"0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624"} Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.142082 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:22Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.142997 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-mhncd" event={"ID":"8dd11f51-430a-4e7a-ab55-9666abfb1e82","Type":"ContainerStarted","Data":"7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809"} Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.160446 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:22Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.184670 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:22Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.200474 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:22Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.233530 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:22Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.250799 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:22Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.266559 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:22Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.379145 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:22Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:22 crc kubenswrapper[4813]: I1206 15:46:22.676091 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:22Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.003449 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:22Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.231558 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:46:23 crc kubenswrapper[4813]: E1206 15:46:23.231702 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:46:27.231678149 +0000 UTC m=+27.122557725 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.231778 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.231821 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.231861 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:46:23 crc kubenswrapper[4813]: E1206 15:46:23.232087 4813 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 15:46:23 crc kubenswrapper[4813]: E1206 15:46:23.232141 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 15:46:27.232130021 +0000 UTC m=+27.123009597 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 15:46:23 crc kubenswrapper[4813]: E1206 15:46:23.232157 4813 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 15:46:23 crc kubenswrapper[4813]: E1206 15:46:23.232237 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 15:46:27.232221453 +0000 UTC m=+27.123101029 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 15:46:23 crc kubenswrapper[4813]: E1206 15:46:23.232333 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 15:46:23 crc kubenswrapper[4813]: E1206 15:46:23.232357 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 15:46:23 crc kubenswrapper[4813]: E1206 15:46:23.232370 4813 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:46:23 crc kubenswrapper[4813]: E1206 15:46:23.232435 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 15:46:27.232418828 +0000 UTC m=+27.123298404 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.236519 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" event={"ID":"0e54e31f-e9f8-4dd5-8738-c785df002c93","Type":"ContainerStarted","Data":"5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c"} Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.238715 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" event={"ID":"6adb3a02-198b-41b4-8b18-56b0f9527fe7","Type":"ContainerStarted","Data":"511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378"} Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.338361 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:23Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.424677 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-frdb4"] Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.424979 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-frdb4" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.427963 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.428251 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.431620 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.431750 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.447402 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:23Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.475344 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:23Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.486544 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:23 crc kubenswrapper[4813]: E1206 15:46:23.486669 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.486726 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:46:23 crc kubenswrapper[4813]: E1206 15:46:23.486785 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.486821 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:46:23 crc kubenswrapper[4813]: E1206 15:46:23.486858 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.509486 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:23Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.536873 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1-host\") pod \"node-ca-frdb4\" (UID: \"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\") " pod="openshift-image-registry/node-ca-frdb4" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.536945 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.536990 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1-serviceca\") pod \"node-ca-frdb4\" (UID: \"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\") " pod="openshift-image-registry/node-ca-frdb4" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.537015 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgn4r\" (UniqueName: \"kubernetes.io/projected/b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1-kube-api-access-vgn4r\") pod \"node-ca-frdb4\" (UID: \"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\") " pod="openshift-image-registry/node-ca-frdb4" Dec 06 15:46:23 crc kubenswrapper[4813]: E1206 15:46:23.537160 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 15:46:23 crc kubenswrapper[4813]: E1206 15:46:23.537187 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 15:46:23 crc kubenswrapper[4813]: E1206 15:46:23.537200 4813 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:46:23 crc kubenswrapper[4813]: E1206 15:46:23.537273 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 15:46:27.537241496 +0000 UTC m=+27.428121072 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.543967 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:23Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.565700 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:23Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.590966 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:23Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.615105 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:23Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.638163 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1-serviceca\") pod \"node-ca-frdb4\" (UID: \"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\") " pod="openshift-image-registry/node-ca-frdb4" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.638202 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgn4r\" (UniqueName: \"kubernetes.io/projected/b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1-kube-api-access-vgn4r\") pod \"node-ca-frdb4\" (UID: \"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\") " pod="openshift-image-registry/node-ca-frdb4" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.638235 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1-host\") pod \"node-ca-frdb4\" (UID: \"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\") " pod="openshift-image-registry/node-ca-frdb4" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.638326 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1-host\") pod \"node-ca-frdb4\" (UID: \"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\") " pod="openshift-image-registry/node-ca-frdb4" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.639448 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1-serviceca\") pod \"node-ca-frdb4\" (UID: \"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\") " pod="openshift-image-registry/node-ca-frdb4" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.648141 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:23Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.679337 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgn4r\" (UniqueName: \"kubernetes.io/projected/b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1-kube-api-access-vgn4r\") pod \"node-ca-frdb4\" (UID: \"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\") " pod="openshift-image-registry/node-ca-frdb4" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.686477 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:23Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.706920 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:23Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.730185 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:23Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.745053 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-frdb4" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.766167 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:23Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.800917 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:23Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.848809 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:23Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:23 crc kubenswrapper[4813]: I1206 15:46:23.892174 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:23Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:24 crc kubenswrapper[4813]: I1206 15:46:24.027220 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:23Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:24 crc kubenswrapper[4813]: I1206 15:46:24.042559 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:24Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:24 crc kubenswrapper[4813]: I1206 15:46:24.101629 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:24Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:24 crc kubenswrapper[4813]: I1206 15:46:24.142842 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:24Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:24 crc kubenswrapper[4813]: I1206 15:46:24.161030 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:24Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:24 crc kubenswrapper[4813]: I1206 15:46:24.177049 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:24Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:24 crc kubenswrapper[4813]: I1206 15:46:24.197876 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:24Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:24 crc kubenswrapper[4813]: I1206 15:46:24.211559 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:24Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:24 crc kubenswrapper[4813]: I1206 15:46:24.243681 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:24Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:24 crc kubenswrapper[4813]: I1206 15:46:24.250426 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-frdb4" event={"ID":"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1","Type":"ContainerStarted","Data":"859e3ffd80a7c7501b26ddf7042aec5e98edb2ec060d678a51fbdef40c14683e"} Dec 06 15:46:24 crc kubenswrapper[4813]: I1206 15:46:24.257169 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" event={"ID":"0e54e31f-e9f8-4dd5-8738-c785df002c93","Type":"ContainerStarted","Data":"2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa"} Dec 06 15:46:24 crc kubenswrapper[4813]: I1206 15:46:24.257206 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" event={"ID":"0e54e31f-e9f8-4dd5-8738-c785df002c93","Type":"ContainerStarted","Data":"6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e"} Dec 06 15:46:24 crc kubenswrapper[4813]: I1206 15:46:24.257216 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" event={"ID":"0e54e31f-e9f8-4dd5-8738-c785df002c93","Type":"ContainerStarted","Data":"f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c"} Dec 06 15:46:24 crc kubenswrapper[4813]: I1206 15:46:24.264744 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:24Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:24 crc kubenswrapper[4813]: I1206 15:46:24.301557 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:24Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:24 crc kubenswrapper[4813]: I1206 15:46:24.316179 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:24Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:24 crc kubenswrapper[4813]: I1206 15:46:24.332998 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:24Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:24 crc kubenswrapper[4813]: I1206 15:46:24.354358 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:24Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.263315 4813 generic.go:334] "Generic (PLEG): container finished" podID="6adb3a02-198b-41b4-8b18-56b0f9527fe7" containerID="511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378" exitCode=0 Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.263390 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" event={"ID":"6adb3a02-198b-41b4-8b18-56b0f9527fe7","Type":"ContainerDied","Data":"511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378"} Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.269050 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" event={"ID":"0e54e31f-e9f8-4dd5-8738-c785df002c93","Type":"ContainerStarted","Data":"2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45"} Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.269078 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" event={"ID":"0e54e31f-e9f8-4dd5-8738-c785df002c93","Type":"ContainerStarted","Data":"57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc"} Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.270987 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-frdb4" event={"ID":"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1","Type":"ContainerStarted","Data":"96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0"} Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.304728 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.331402 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.358121 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.375068 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.394816 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.422592 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.442816 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.459580 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.478721 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.488574 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:25 crc kubenswrapper[4813]: E1206 15:46:25.488736 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.488819 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:46:25 crc kubenswrapper[4813]: E1206 15:46:25.488876 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.488920 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:46:25 crc kubenswrapper[4813]: E1206 15:46:25.488981 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.493712 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.513904 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.537989 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.552203 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.566664 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.582632 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.604000 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.621066 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.637383 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.650869 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.680166 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.710926 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.712721 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.712770 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.712785 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.712906 4813 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.725568 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.725931 4813 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.726161 4813 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.727249 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.727317 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.727329 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.727352 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.727364 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:25Z","lastTransitionTime":"2025-12-06T15:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:25 crc kubenswrapper[4813]: E1206 15:46:25.751293 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.758117 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.758180 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.758192 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.758213 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.758224 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:25Z","lastTransitionTime":"2025-12-06T15:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.758521 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: E1206 15:46:25.772886 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.774209 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.777338 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.777367 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.777378 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.777397 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.777408 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:25Z","lastTransitionTime":"2025-12-06T15:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.789334 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: E1206 15:46:25.790433 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.794113 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.794175 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.794192 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.794232 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.794276 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:25Z","lastTransitionTime":"2025-12-06T15:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.802857 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: E1206 15:46:25.805219 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.808688 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.808744 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.808779 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.808799 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.808812 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:25Z","lastTransitionTime":"2025-12-06T15:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.818857 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: E1206 15:46:25.821430 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: E1206 15:46:25.821627 4813 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.823531 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.823578 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.823596 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.823616 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.823629 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:25Z","lastTransitionTime":"2025-12-06T15:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.831298 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.845460 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.857628 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.871807 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:25Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.925984 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.926226 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.926331 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.926415 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:25 crc kubenswrapper[4813]: I1206 15:46:25.926475 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:25Z","lastTransitionTime":"2025-12-06T15:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.029074 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.029117 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.029125 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.029141 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.029150 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:26Z","lastTransitionTime":"2025-12-06T15:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.135839 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.135879 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.135889 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.135905 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.135916 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:26Z","lastTransitionTime":"2025-12-06T15:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.239468 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.239541 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.239564 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.239592 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.239615 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:26Z","lastTransitionTime":"2025-12-06T15:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.278879 4813 generic.go:334] "Generic (PLEG): container finished" podID="6adb3a02-198b-41b4-8b18-56b0f9527fe7" containerID="87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff" exitCode=0 Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.279023 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" event={"ID":"6adb3a02-198b-41b4-8b18-56b0f9527fe7","Type":"ContainerDied","Data":"87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff"} Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.304955 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:26Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.327337 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:26Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.342174 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.342213 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.342225 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.342242 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.342254 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:26Z","lastTransitionTime":"2025-12-06T15:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.349406 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:26Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.364111 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:26Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.379298 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:26Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.390292 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:26Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.412721 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:26Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.425187 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:26Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.445292 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.445325 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.445338 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.448394 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:26Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.448470 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.448492 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:26Z","lastTransitionTime":"2025-12-06T15:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.472095 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:26Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.487682 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:26Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.500980 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:26Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.514036 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:26Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.527541 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:26Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.548143 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:26Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.549865 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.549905 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.549918 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.549934 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.549947 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:26Z","lastTransitionTime":"2025-12-06T15:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.651397 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.651426 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.651435 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.651447 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.651455 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:26Z","lastTransitionTime":"2025-12-06T15:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.753964 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.754001 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.754013 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.754030 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.754042 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:26Z","lastTransitionTime":"2025-12-06T15:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.856006 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.856033 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.856042 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.856055 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.856064 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:26Z","lastTransitionTime":"2025-12-06T15:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.958588 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.958625 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.958647 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.958662 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:26 crc kubenswrapper[4813]: I1206 15:46:26.958672 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:26Z","lastTransitionTime":"2025-12-06T15:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.060669 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.060699 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.060708 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.060722 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.060730 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:27Z","lastTransitionTime":"2025-12-06T15:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.162819 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.162851 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.162859 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.162872 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.162882 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:27Z","lastTransitionTime":"2025-12-06T15:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.235941 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.236123 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:27 crc kubenswrapper[4813]: E1206 15:46:27.236191 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:46:35.236152909 +0000 UTC m=+35.127032505 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:46:27 crc kubenswrapper[4813]: E1206 15:46:27.236299 4813 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 15:46:27 crc kubenswrapper[4813]: E1206 15:46:27.236377 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 15:46:35.236354125 +0000 UTC m=+35.127233741 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 15:46:27 crc kubenswrapper[4813]: E1206 15:46:27.236389 4813 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 15:46:27 crc kubenswrapper[4813]: E1206 15:46:27.236448 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 15:46:35.236435387 +0000 UTC m=+35.127315053 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.236289 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.236549 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:46:27 crc kubenswrapper[4813]: E1206 15:46:27.236735 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 15:46:27 crc kubenswrapper[4813]: E1206 15:46:27.236781 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 15:46:27 crc kubenswrapper[4813]: E1206 15:46:27.236804 4813 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:46:27 crc kubenswrapper[4813]: E1206 15:46:27.236898 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 15:46:35.236862168 +0000 UTC m=+35.127741824 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.265791 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.265843 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.265859 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.265881 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.265900 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:27Z","lastTransitionTime":"2025-12-06T15:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.287749 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" event={"ID":"0e54e31f-e9f8-4dd5-8738-c785df002c93","Type":"ContainerStarted","Data":"afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2"} Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.290920 4813 generic.go:334] "Generic (PLEG): container finished" podID="6adb3a02-198b-41b4-8b18-56b0f9527fe7" containerID="5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548" exitCode=0 Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.290970 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" event={"ID":"6adb3a02-198b-41b4-8b18-56b0f9527fe7","Type":"ContainerDied","Data":"5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548"} Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.323535 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:27Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.345506 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:27Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.365082 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:27Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.370357 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.370399 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.370413 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.370434 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.370450 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:27Z","lastTransitionTime":"2025-12-06T15:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.382677 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:27Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.406052 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:27Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.417653 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:27Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.428002 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:27Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.458465 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:27Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.472475 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.472514 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.472526 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.472542 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.472552 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:27Z","lastTransitionTime":"2025-12-06T15:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.476812 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:27Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.486241 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.486339 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:27 crc kubenswrapper[4813]: E1206 15:46:27.486476 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:46:27 crc kubenswrapper[4813]: E1206 15:46:27.486391 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.486541 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:46:27 crc kubenswrapper[4813]: E1206 15:46:27.486592 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.488854 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:27Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.503203 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:27Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.515189 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:27Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.526085 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:27Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.538759 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:27Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.539205 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:46:27 crc kubenswrapper[4813]: E1206 15:46:27.539353 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 15:46:27 crc kubenswrapper[4813]: E1206 15:46:27.539367 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 15:46:27 crc kubenswrapper[4813]: E1206 15:46:27.539377 4813 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:46:27 crc kubenswrapper[4813]: E1206 15:46:27.539412 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 15:46:35.539399787 +0000 UTC m=+35.430279353 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.552520 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:27Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.575708 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.575742 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.575755 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.575770 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.575781 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:27Z","lastTransitionTime":"2025-12-06T15:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.678532 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.678795 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.678807 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.678827 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.678841 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:27Z","lastTransitionTime":"2025-12-06T15:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.781908 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.781966 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.781985 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.782010 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.782027 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:27Z","lastTransitionTime":"2025-12-06T15:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.885015 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.885061 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.885078 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.885120 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.885137 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:27Z","lastTransitionTime":"2025-12-06T15:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.988421 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.988459 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.988471 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.988488 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:27 crc kubenswrapper[4813]: I1206 15:46:27.988499 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:27Z","lastTransitionTime":"2025-12-06T15:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.090728 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.090784 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.090800 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.090830 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.090849 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:28Z","lastTransitionTime":"2025-12-06T15:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.193139 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.193189 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.193201 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.193221 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.193235 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:28Z","lastTransitionTime":"2025-12-06T15:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.295039 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.295079 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.295089 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.295104 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.295113 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:28Z","lastTransitionTime":"2025-12-06T15:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.296818 4813 generic.go:334] "Generic (PLEG): container finished" podID="6adb3a02-198b-41b4-8b18-56b0f9527fe7" containerID="059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903" exitCode=0 Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.296863 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" event={"ID":"6adb3a02-198b-41b4-8b18-56b0f9527fe7","Type":"ContainerDied","Data":"059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903"} Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.315810 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:28Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.331620 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:28Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.347307 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:28Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.362969 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:28Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.380616 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:28Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.398730 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.398778 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.398794 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.398822 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.398838 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:28Z","lastTransitionTime":"2025-12-06T15:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.404170 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:28Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.419157 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:28Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.430331 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:28Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.454795 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:28Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.469658 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:28Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.482313 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:28Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.503403 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.503434 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.503443 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.503463 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.503473 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:28Z","lastTransitionTime":"2025-12-06T15:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.504768 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:28Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.520784 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:28Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.532410 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:28Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.547549 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:28Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.605881 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.605922 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.605935 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.605951 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.605962 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:28Z","lastTransitionTime":"2025-12-06T15:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.708375 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.708405 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.708415 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.708430 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.708442 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:28Z","lastTransitionTime":"2025-12-06T15:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.811667 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.811725 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.811737 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.811759 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.811770 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:28Z","lastTransitionTime":"2025-12-06T15:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.914206 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.914240 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.914249 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.914280 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:28 crc kubenswrapper[4813]: I1206 15:46:28.914295 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:28Z","lastTransitionTime":"2025-12-06T15:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.016058 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.016079 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.016086 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.016098 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.016106 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:29Z","lastTransitionTime":"2025-12-06T15:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.118339 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.118396 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.118417 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.118445 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.118461 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:29Z","lastTransitionTime":"2025-12-06T15:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.221800 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.221846 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.221865 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.221888 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.221906 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:29Z","lastTransitionTime":"2025-12-06T15:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.304620 4813 generic.go:334] "Generic (PLEG): container finished" podID="6adb3a02-198b-41b4-8b18-56b0f9527fe7" containerID="8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82" exitCode=0 Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.304658 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" event={"ID":"6adb3a02-198b-41b4-8b18-56b0f9527fe7","Type":"ContainerDied","Data":"8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82"} Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.317894 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" event={"ID":"0e54e31f-e9f8-4dd5-8738-c785df002c93","Type":"ContainerStarted","Data":"35430fe750629826f2ce9655a41ac378ba7f18528952fa40a2557faf1b9c5743"} Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.318595 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.318639 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.325628 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.325672 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.325690 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.325713 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.325729 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:29Z","lastTransitionTime":"2025-12-06T15:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.329424 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:29Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.353939 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.355647 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:29Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.358559 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.369931 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:29Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.393463 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:29Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.405653 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:29Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.426970 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:29Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.428137 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.428187 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.428204 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.428237 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.428440 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:29Z","lastTransitionTime":"2025-12-06T15:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.439366 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:29Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.453107 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:29Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.464886 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:29Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.476704 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:29Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.486148 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:29Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.486292 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.486366 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.486455 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:46:29 crc kubenswrapper[4813]: E1206 15:46:29.486451 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:46:29 crc kubenswrapper[4813]: E1206 15:46:29.486494 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:46:29 crc kubenswrapper[4813]: E1206 15:46:29.486581 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.503783 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:29Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.512971 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:29Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.522864 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:29Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.531731 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.531781 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.531795 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.531812 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.531826 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:29Z","lastTransitionTime":"2025-12-06T15:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.535574 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:29Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.547855 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:29Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.558816 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:29Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.571752 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:29Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.586367 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:29Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.597110 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:29Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.613228 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:29Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.625120 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:29Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.634601 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.634635 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.634644 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.634658 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.634667 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:29Z","lastTransitionTime":"2025-12-06T15:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.640555 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:29Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.656430 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35430fe750629826f2ce9655a41ac378ba7f18528952fa40a2557faf1b9c5743\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:29Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.664678 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:29Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.675815 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:29Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.684943 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:29Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.694235 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:29Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.703978 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:29Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.715252 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:29Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.737192 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.737404 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.737526 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.737636 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.737749 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:29Z","lastTransitionTime":"2025-12-06T15:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.841350 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.841386 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.841398 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.841415 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.841427 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:29Z","lastTransitionTime":"2025-12-06T15:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.944559 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.944885 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.945016 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.945136 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:29 crc kubenswrapper[4813]: I1206 15:46:29.945252 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:29Z","lastTransitionTime":"2025-12-06T15:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.048033 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.048074 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.048087 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.048108 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.048123 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:30Z","lastTransitionTime":"2025-12-06T15:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.151468 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.151506 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.151520 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.151536 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.151547 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:30Z","lastTransitionTime":"2025-12-06T15:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.255477 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.255514 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.255527 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.255545 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.255555 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:30Z","lastTransitionTime":"2025-12-06T15:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.329326 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.330028 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" event={"ID":"6adb3a02-198b-41b4-8b18-56b0f9527fe7","Type":"ContainerStarted","Data":"f228acd2a1febab7d6991054e11cf509ff83ab26e537db82ab49688144a4a233"} Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.354335 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.358607 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.358643 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.358658 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.358679 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.358692 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:30Z","lastTransitionTime":"2025-12-06T15:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.368200 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.390444 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.407876 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.426373 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.447075 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f228acd2a1febab7d6991054e11cf509ff83ab26e537db82ab49688144a4a233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.461536 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.461604 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.461624 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.461649 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.461666 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:30Z","lastTransitionTime":"2025-12-06T15:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.464209 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.491683 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.507747 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.527050 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.541770 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.564780 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.564815 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.564828 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.564844 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.564856 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:30Z","lastTransitionTime":"2025-12-06T15:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.565895 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.584056 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.607128 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35430fe750629826f2ce9655a41ac378ba7f18528952fa40a2557faf1b9c5743\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.627038 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.648180 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.667218 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.667247 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.667269 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.667283 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.667292 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:30Z","lastTransitionTime":"2025-12-06T15:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.668030 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.682612 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.696844 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.710686 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.722894 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.732902 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.749555 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35430fe750629826f2ce9655a41ac378ba7f18528952fa40a2557faf1b9c5743\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.764435 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.768970 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.769013 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.769026 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.769045 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.769058 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:30Z","lastTransitionTime":"2025-12-06T15:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.780656 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.792050 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.804719 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.816290 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.835110 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f228acd2a1febab7d6991054e11cf509ff83ab26e537db82ab49688144a4a233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.851603 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.857831 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.872520 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.872680 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.872798 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.872889 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.872952 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:30Z","lastTransitionTime":"2025-12-06T15:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.981068 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.981156 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.981169 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.981185 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:30 crc kubenswrapper[4813]: I1206 15:46:30.981195 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:30Z","lastTransitionTime":"2025-12-06T15:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.083321 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.083348 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.083359 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.083375 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.083387 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:31Z","lastTransitionTime":"2025-12-06T15:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.190017 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.190057 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.190067 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.190082 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.190092 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:31Z","lastTransitionTime":"2025-12-06T15:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.292192 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.292224 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.292237 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.292252 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.292281 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:31Z","lastTransitionTime":"2025-12-06T15:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.332301 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"b0a379c66da38cd6d16eb0fdc5941e3fd5d4127b7cce84fe2bde55918ef2c99c"} Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.352396 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:31Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.363098 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:31Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.380514 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.384383 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a379c66da38cd6d16eb0fdc5941e3fd5d4127b7cce84fe2bde55918ef2c99c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:31Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.400125 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f228acd2a1febab7d6991054e11cf509ff83ab26e537db82ab49688144a4a233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:31Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.413593 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:31Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.414580 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.414610 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.414620 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.414634 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.414642 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:31Z","lastTransitionTime":"2025-12-06T15:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.428690 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:31Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.441411 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:31Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.460615 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:31Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.474773 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:31Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.486325 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.486364 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.486373 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:46:31 crc kubenswrapper[4813]: E1206 15:46:31.486506 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:46:31 crc kubenswrapper[4813]: E1206 15:46:31.486637 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:46:31 crc kubenswrapper[4813]: E1206 15:46:31.486725 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.489468 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:31Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.500339 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:31Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.517109 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.517146 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.517156 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.517173 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.517183 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:31Z","lastTransitionTime":"2025-12-06T15:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.521640 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35430fe750629826f2ce9655a41ac378ba7f18528952fa40a2557faf1b9c5743\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:31Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.536066 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:31Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.549639 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:31Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.571328 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:31Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.584879 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:31Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.596723 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:31Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.610461 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:31Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.618895 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.618932 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.618949 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.618966 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.618977 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:31Z","lastTransitionTime":"2025-12-06T15:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.625518 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:31Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.637305 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a379c66da38cd6d16eb0fdc5941e3fd5d4127b7cce84fe2bde55918ef2c99c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:31Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.654214 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f228acd2a1febab7d6991054e11cf509ff83ab26e537db82ab49688144a4a233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:31Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.668830 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:31Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.692133 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:31Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.709129 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:31Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.721498 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.721540 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.721551 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.721569 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.721580 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:31Z","lastTransitionTime":"2025-12-06T15:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.723639 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:31Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.737990 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:31Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.758215 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:31Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.770894 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:31Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.789276 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35430fe750629826f2ce9655a41ac378ba7f18528952fa40a2557faf1b9c5743\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:31Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.799240 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:31Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.832117 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.832155 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.832180 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.832198 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.832210 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:31Z","lastTransitionTime":"2025-12-06T15:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.934576 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.934611 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.934623 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.934640 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:31 crc kubenswrapper[4813]: I1206 15:46:31.934653 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:31Z","lastTransitionTime":"2025-12-06T15:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.038510 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.038830 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.038976 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.039151 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.039304 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:32Z","lastTransitionTime":"2025-12-06T15:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.142454 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.142517 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.142534 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.142559 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.142577 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:32Z","lastTransitionTime":"2025-12-06T15:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.245747 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.245796 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.245813 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.245837 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.245879 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:32Z","lastTransitionTime":"2025-12-06T15:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.348755 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.348813 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.348840 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.348866 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.348887 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:32Z","lastTransitionTime":"2025-12-06T15:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.452045 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.452104 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.452121 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.452144 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.452160 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:32Z","lastTransitionTime":"2025-12-06T15:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.555759 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.555837 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.555855 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.555877 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.555895 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:32Z","lastTransitionTime":"2025-12-06T15:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.658748 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.659091 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.659229 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.659488 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.659653 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:32Z","lastTransitionTime":"2025-12-06T15:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.762635 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.763012 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.763209 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.763380 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.763517 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:32Z","lastTransitionTime":"2025-12-06T15:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.866409 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.866451 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.866467 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.866516 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.866535 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:32Z","lastTransitionTime":"2025-12-06T15:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.969811 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.970138 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.970310 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.970477 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:32 crc kubenswrapper[4813]: I1206 15:46:32.970650 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:32Z","lastTransitionTime":"2025-12-06T15:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.074374 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.074426 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.074443 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.074469 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.074486 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:33Z","lastTransitionTime":"2025-12-06T15:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.178082 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.178147 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.178164 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.178188 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.178209 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:33Z","lastTransitionTime":"2025-12-06T15:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.281606 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.281673 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.281690 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.281716 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.281738 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:33Z","lastTransitionTime":"2025-12-06T15:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.342499 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xdt4d_0e54e31f-e9f8-4dd5-8738-c785df002c93/ovnkube-controller/0.log" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.348633 4813 generic.go:334] "Generic (PLEG): container finished" podID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerID="35430fe750629826f2ce9655a41ac378ba7f18528952fa40a2557faf1b9c5743" exitCode=1 Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.348696 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" event={"ID":"0e54e31f-e9f8-4dd5-8738-c785df002c93","Type":"ContainerDied","Data":"35430fe750629826f2ce9655a41ac378ba7f18528952fa40a2557faf1b9c5743"} Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.349932 4813 scope.go:117] "RemoveContainer" containerID="35430fe750629826f2ce9655a41ac378ba7f18528952fa40a2557faf1b9c5743" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.374748 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:33Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.385497 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.385561 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.385587 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.385618 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.385641 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:33Z","lastTransitionTime":"2025-12-06T15:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.395316 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:33Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.417530 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:33Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.435289 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:33Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.449956 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a379c66da38cd6d16eb0fdc5941e3fd5d4127b7cce84fe2bde55918ef2c99c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:33Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.474063 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f228acd2a1febab7d6991054e11cf509ff83ab26e537db82ab49688144a4a233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:33Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.486251 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.486330 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.486251 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:33 crc kubenswrapper[4813]: E1206 15:46:33.486671 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:46:33 crc kubenswrapper[4813]: E1206 15:46:33.486840 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:46:33 crc kubenswrapper[4813]: E1206 15:46:33.486943 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.490442 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:33Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.492093 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.492141 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.492159 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.492181 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.492199 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:33Z","lastTransitionTime":"2025-12-06T15:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.518033 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:33Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.530661 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:33Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.547304 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:33Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.566479 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:33Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.583204 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:33Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.595859 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.595918 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.595944 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.595974 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.595998 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:33Z","lastTransitionTime":"2025-12-06T15:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.605110 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:33Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.618926 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:33Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.643623 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35430fe750629826f2ce9655a41ac378ba7f18528952fa40a2557faf1b9c5743\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35430fe750629826f2ce9655a41ac378ba7f18528952fa40a2557faf1b9c5743\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T15:46:32Z\\\",\\\"message\\\":\\\"9042 5996 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 15:46:32.039137 5996 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 15:46:32.039201 5996 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 15:46:32.039298 5996 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 15:46:32.039508 5996 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 15:46:32.039687 5996 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 15:46:32.039709 5996 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 15:46:32.039743 5996 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 15:46:32.040362 5996 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 15:46:32.040486 5996 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 15:46:32.040604 5996 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 15:46:32.040809 5996 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:33Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.699389 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.699445 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.699464 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.699487 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.699504 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:33Z","lastTransitionTime":"2025-12-06T15:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.802376 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.802428 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.802442 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.802461 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.802471 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:33Z","lastTransitionTime":"2025-12-06T15:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.905913 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.906291 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.906302 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.906316 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:33 crc kubenswrapper[4813]: I1206 15:46:33.906328 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:33Z","lastTransitionTime":"2025-12-06T15:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.008775 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.008803 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.008811 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.008826 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.008836 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:34Z","lastTransitionTime":"2025-12-06T15:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.065976 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr"] Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.066436 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.069006 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.069068 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.085181 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.097293 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.109183 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.110494 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.110537 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.110549 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.110565 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.110573 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:34Z","lastTransitionTime":"2025-12-06T15:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.124487 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.137059 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.149217 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a379c66da38cd6d16eb0fdc5941e3fd5d4127b7cce84fe2bde55918ef2c99c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.149568 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b9de44d2-42dc-46fb-9780-4dcd9e5f9c86-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bd4fr\" (UID: \"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.149598 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvwp6\" (UniqueName: \"kubernetes.io/projected/b9de44d2-42dc-46fb-9780-4dcd9e5f9c86-kube-api-access-wvwp6\") pod \"ovnkube-control-plane-749d76644c-bd4fr\" (UID: \"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.149735 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b9de44d2-42dc-46fb-9780-4dcd9e5f9c86-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bd4fr\" (UID: \"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.149763 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b9de44d2-42dc-46fb-9780-4dcd9e5f9c86-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bd4fr\" (UID: \"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.166849 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f228acd2a1febab7d6991054e11cf509ff83ab26e537db82ab49688144a4a233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.195122 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.207704 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bd4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.212973 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.213012 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.213021 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.213036 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.213047 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:34Z","lastTransitionTime":"2025-12-06T15:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.219146 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.236367 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35430fe750629826f2ce9655a41ac378ba7f18528952fa40a2557faf1b9c5743\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35430fe750629826f2ce9655a41ac378ba7f18528952fa40a2557faf1b9c5743\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T15:46:32Z\\\",\\\"message\\\":\\\"9042 5996 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 15:46:32.039137 5996 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 15:46:32.039201 5996 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 15:46:32.039298 5996 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 15:46:32.039508 5996 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 15:46:32.039687 5996 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 15:46:32.039709 5996 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 15:46:32.039743 5996 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 15:46:32.040362 5996 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 15:46:32.040486 5996 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 15:46:32.040604 5996 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 15:46:32.040809 5996 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.246298 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.250279 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b9de44d2-42dc-46fb-9780-4dcd9e5f9c86-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bd4fr\" (UID: \"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.250327 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvwp6\" (UniqueName: \"kubernetes.io/projected/b9de44d2-42dc-46fb-9780-4dcd9e5f9c86-kube-api-access-wvwp6\") pod \"ovnkube-control-plane-749d76644c-bd4fr\" (UID: \"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.250383 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b9de44d2-42dc-46fb-9780-4dcd9e5f9c86-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bd4fr\" (UID: \"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.250419 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b9de44d2-42dc-46fb-9780-4dcd9e5f9c86-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bd4fr\" (UID: \"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.253495 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b9de44d2-42dc-46fb-9780-4dcd9e5f9c86-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bd4fr\" (UID: \"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.253526 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b9de44d2-42dc-46fb-9780-4dcd9e5f9c86-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bd4fr\" (UID: \"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.258653 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b9de44d2-42dc-46fb-9780-4dcd9e5f9c86-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bd4fr\" (UID: \"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.267828 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.275534 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvwp6\" (UniqueName: \"kubernetes.io/projected/b9de44d2-42dc-46fb-9780-4dcd9e5f9c86-kube-api-access-wvwp6\") pod \"ovnkube-control-plane-749d76644c-bd4fr\" (UID: \"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.280451 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.292249 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.304003 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.315614 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.315652 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.315664 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.315683 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.315694 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:34Z","lastTransitionTime":"2025-12-06T15:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.353597 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xdt4d_0e54e31f-e9f8-4dd5-8738-c785df002c93/ovnkube-controller/0.log" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.356360 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" event={"ID":"0e54e31f-e9f8-4dd5-8738-c785df002c93","Type":"ContainerStarted","Data":"ccdd42da55ee2466605cb3f7dd612f6da8e15ec9f5305f7dfd06d0928aad263f"} Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.357193 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.371496 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.379412 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.384804 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.401710 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: W1206 15:46:34.402586 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9de44d2_42dc_46fb_9780_4dcd9e5f9c86.slice/crio-d7038e0ba307d7fcde28bc10170d7eb76181ff287bafc1f6fc28fc7ded3b047e WatchSource:0}: Error finding container d7038e0ba307d7fcde28bc10170d7eb76181ff287bafc1f6fc28fc7ded3b047e: Status 404 returned error can't find the container with id d7038e0ba307d7fcde28bc10170d7eb76181ff287bafc1f6fc28fc7ded3b047e Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.415058 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.419441 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.419480 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.419489 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.419506 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.419515 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:34Z","lastTransitionTime":"2025-12-06T15:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.427657 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.458542 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdd42da55ee2466605cb3f7dd612f6da8e15ec9f5305f7dfd06d0928aad263f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35430fe750629826f2ce9655a41ac378ba7f18528952fa40a2557faf1b9c5743\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T15:46:32Z\\\",\\\"message\\\":\\\"9042 5996 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 15:46:32.039137 5996 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 15:46:32.039201 5996 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 15:46:32.039298 5996 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 15:46:32.039508 5996 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 15:46:32.039687 5996 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 15:46:32.039709 5996 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 15:46:32.039743 5996 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 15:46:32.040362 5996 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 15:46:32.040486 5996 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 15:46:32.040604 5996 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 15:46:32.040809 5996 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.467428 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.478646 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.489346 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.501778 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.512707 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.522303 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.522343 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.522354 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.522370 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.522381 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:34Z","lastTransitionTime":"2025-12-06T15:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.523211 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a379c66da38cd6d16eb0fdc5941e3fd5d4127b7cce84fe2bde55918ef2c99c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.536963 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f228acd2a1febab7d6991054e11cf509ff83ab26e537db82ab49688144a4a233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.550010 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.568827 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.579783 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bd4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.624819 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.624877 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.624895 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.624923 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.624941 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:34Z","lastTransitionTime":"2025-12-06T15:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.727253 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.727315 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.727325 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.727345 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.727356 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:34Z","lastTransitionTime":"2025-12-06T15:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.832976 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.833037 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.833055 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.833082 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.833099 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:34Z","lastTransitionTime":"2025-12-06T15:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.935587 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.935643 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.935665 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.935689 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:34 crc kubenswrapper[4813]: I1206 15:46:34.935708 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:34Z","lastTransitionTime":"2025-12-06T15:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.038773 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.038847 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.038862 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.038890 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.038905 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:35Z","lastTransitionTime":"2025-12-06T15:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.142131 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.142186 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.142197 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.142212 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.142223 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:35Z","lastTransitionTime":"2025-12-06T15:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.244708 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.244778 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.244802 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.244829 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.244864 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:35Z","lastTransitionTime":"2025-12-06T15:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.259467 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.259557 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.259591 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.259626 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:46:35 crc kubenswrapper[4813]: E1206 15:46:35.259763 4813 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 15:46:35 crc kubenswrapper[4813]: E1206 15:46:35.259767 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 15:46:35 crc kubenswrapper[4813]: E1206 15:46:35.259861 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 15:46:51.259824667 +0000 UTC m=+51.150704253 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 15:46:35 crc kubenswrapper[4813]: E1206 15:46:35.259907 4813 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 15:46:35 crc kubenswrapper[4813]: E1206 15:46:35.259931 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:46:51.259875889 +0000 UTC m=+51.150755505 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:46:35 crc kubenswrapper[4813]: E1206 15:46:35.260028 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 15:46:51.259991242 +0000 UTC m=+51.150870868 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 15:46:35 crc kubenswrapper[4813]: E1206 15:46:35.259910 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 15:46:35 crc kubenswrapper[4813]: E1206 15:46:35.260097 4813 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:46:35 crc kubenswrapper[4813]: E1206 15:46:35.260188 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 15:46:51.260157956 +0000 UTC m=+51.151037572 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.348612 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.348729 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.348753 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.348783 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.348808 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:35Z","lastTransitionTime":"2025-12-06T15:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.363505 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" event={"ID":"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86","Type":"ContainerStarted","Data":"b14a8cf06c49b6922327f25905466f86e1a0fd0a0b852f751bac6f8cd7db9e4a"} Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.363598 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" event={"ID":"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86","Type":"ContainerStarted","Data":"3ab1d9f15bba233c4bbd04ece144a5a8119c168abaeee7d4544269b55ad58ddb"} Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.363626 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" event={"ID":"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86","Type":"ContainerStarted","Data":"d7038e0ba307d7fcde28bc10170d7eb76181ff287bafc1f6fc28fc7ded3b047e"} Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.366364 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xdt4d_0e54e31f-e9f8-4dd5-8738-c785df002c93/ovnkube-controller/1.log" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.367187 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xdt4d_0e54e31f-e9f8-4dd5-8738-c785df002c93/ovnkube-controller/0.log" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.372299 4813 generic.go:334] "Generic (PLEG): container finished" podID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerID="ccdd42da55ee2466605cb3f7dd612f6da8e15ec9f5305f7dfd06d0928aad263f" exitCode=2 Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.372351 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" event={"ID":"0e54e31f-e9f8-4dd5-8738-c785df002c93","Type":"ContainerDied","Data":"ccdd42da55ee2466605cb3f7dd612f6da8e15ec9f5305f7dfd06d0928aad263f"} Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.372417 4813 scope.go:117] "RemoveContainer" containerID="35430fe750629826f2ce9655a41ac378ba7f18528952fa40a2557faf1b9c5743" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.373454 4813 scope.go:117] "RemoveContainer" containerID="ccdd42da55ee2466605cb3f7dd612f6da8e15ec9f5305f7dfd06d0928aad263f" Dec 06 15:46:35 crc kubenswrapper[4813]: E1206 15:46:35.373702 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-xdt4d_openshift-ovn-kubernetes(0e54e31f-e9f8-4dd5-8738-c785df002c93)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.390585 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.404591 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.427949 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.446616 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.452052 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.452114 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.452203 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.452242 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.452315 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:35Z","lastTransitionTime":"2025-12-06T15:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.463362 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a379c66da38cd6d16eb0fdc5941e3fd5d4127b7cce84fe2bde55918ef2c99c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.486561 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.486661 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.486992 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:35 crc kubenswrapper[4813]: E1206 15:46:35.487186 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:46:35 crc kubenswrapper[4813]: E1206 15:46:35.487364 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:46:35 crc kubenswrapper[4813]: E1206 15:46:35.487510 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.488145 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f228acd2a1febab7d6991054e11cf509ff83ab26e537db82ab49688144a4a233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.510725 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.534642 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.551207 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ab1d9f15bba233c4bbd04ece144a5a8119c168abaeee7d4544269b55ad58ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b14a8cf06c49b6922327f25905466f86e1a0fd0a0b852f751bac6f8cd7db9e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bd4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.553218 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-kzbhb"] Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.554024 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:46:35 crc kubenswrapper[4813]: E1206 15:46:35.554114 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.554945 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.554983 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.554995 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.555020 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.555031 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:35Z","lastTransitionTime":"2025-12-06T15:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.563244 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18a94103-5ed7-4e27-a58d-7c989b0f70f3-metrics-certs\") pod \"network-metrics-daemon-kzbhb\" (UID: \"18a94103-5ed7-4e27-a58d-7c989b0f70f3\") " pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.563386 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8s4r\" (UniqueName: \"kubernetes.io/projected/18a94103-5ed7-4e27-a58d-7c989b0f70f3-kube-api-access-d8s4r\") pod \"network-metrics-daemon-kzbhb\" (UID: \"18a94103-5ed7-4e27-a58d-7c989b0f70f3\") " pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.563497 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:46:35 crc kubenswrapper[4813]: E1206 15:46:35.563689 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 15:46:35 crc kubenswrapper[4813]: E1206 15:46:35.563749 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 15:46:35 crc kubenswrapper[4813]: E1206 15:46:35.563770 4813 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:46:35 crc kubenswrapper[4813]: E1206 15:46:35.563849 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 15:46:51.563823274 +0000 UTC m=+51.454702890 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.572560 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.588062 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.602830 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.620706 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.635405 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.658205 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.658627 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.658653 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.658687 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.658710 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:35Z","lastTransitionTime":"2025-12-06T15:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.664786 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18a94103-5ed7-4e27-a58d-7c989b0f70f3-metrics-certs\") pod \"network-metrics-daemon-kzbhb\" (UID: \"18a94103-5ed7-4e27-a58d-7c989b0f70f3\") " pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.664855 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8s4r\" (UniqueName: \"kubernetes.io/projected/18a94103-5ed7-4e27-a58d-7c989b0f70f3-kube-api-access-d8s4r\") pod \"network-metrics-daemon-kzbhb\" (UID: \"18a94103-5ed7-4e27-a58d-7c989b0f70f3\") " pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:46:35 crc kubenswrapper[4813]: E1206 15:46:35.664984 4813 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 15:46:35 crc kubenswrapper[4813]: E1206 15:46:35.665097 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18a94103-5ed7-4e27-a58d-7c989b0f70f3-metrics-certs podName:18a94103-5ed7-4e27-a58d-7c989b0f70f3 nodeName:}" failed. No retries permitted until 2025-12-06 15:46:36.165065847 +0000 UTC m=+36.055945473 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/18a94103-5ed7-4e27-a58d-7c989b0f70f3-metrics-certs") pod "network-metrics-daemon-kzbhb" (UID: "18a94103-5ed7-4e27-a58d-7c989b0f70f3") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.667530 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdd42da55ee2466605cb3f7dd612f6da8e15ec9f5305f7dfd06d0928aad263f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35430fe750629826f2ce9655a41ac378ba7f18528952fa40a2557faf1b9c5743\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T15:46:32Z\\\",\\\"message\\\":\\\"9042 5996 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 15:46:32.039137 5996 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 15:46:32.039201 5996 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 15:46:32.039298 5996 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 15:46:32.039508 5996 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 15:46:32.039687 5996 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 15:46:32.039709 5996 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 15:46:32.039743 5996 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 15:46:32.040362 5996 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 15:46:32.040486 5996 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 15:46:32.040604 5996 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 15:46:32.040809 5996 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.682583 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.687513 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8s4r\" (UniqueName: \"kubernetes.io/projected/18a94103-5ed7-4e27-a58d-7c989b0f70f3-kube-api-access-d8s4r\") pod \"network-metrics-daemon-kzbhb\" (UID: \"18a94103-5ed7-4e27-a58d-7c989b0f70f3\") " pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.698335 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.712291 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.731278 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.748931 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.760909 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.760973 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.760990 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.761014 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.761032 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:35Z","lastTransitionTime":"2025-12-06T15:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.764473 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.778546 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.803814 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdd42da55ee2466605cb3f7dd612f6da8e15ec9f5305f7dfd06d0928aad263f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35430fe750629826f2ce9655a41ac378ba7f18528952fa40a2557faf1b9c5743\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T15:46:32Z\\\",\\\"message\\\":\\\"9042 5996 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 15:46:32.039137 5996 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 15:46:32.039201 5996 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 15:46:32.039298 5996 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 15:46:32.039508 5996 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 15:46:32.039687 5996 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 15:46:32.039709 5996 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 15:46:32.039743 5996 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 15:46:32.040362 5996 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 15:46:32.040486 5996 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 15:46:32.040604 5996 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 15:46:32.040809 5996 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccdd42da55ee2466605cb3f7dd612f6da8e15ec9f5305f7dfd06d0928aad263f\\\",\\\"exitCode\\\":2,\\\"finishedAt\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"message\\\":\\\"/127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z\\\\nI1206 15:46:34.429939 6139 nad_controller.go:166] [node-nad-controller NAD controller]: shutting down\\\\nI1206 15:46:34.429795 6139 lb_config.go:1031] Cluster endpoints for openshift-kube-storage-version-migrator-operator/metrics for network=default are: map[]\\\\nI1206 15:46:34.429993 6139 services_controller.go:443] Built service openshift-kube-storage-version-migrator-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.36\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1206 15:46:34.429986 6139 udn_isolation.go:361] D-Bus event received: \\\\u0026dbus.Signal{Sender:\\\\\\\"org.freedesktop.systemd1\\\\\\\", Path:\\\\\\\"/org/freedesktop/systemd1\\\\\\\", Name:\\\\\\\"org.freedesktop.systemd1.Manager.UnitNew\\\\\\\", Body:[]interface {}{\\\\\\\"var-lib-containers.mount\\\\\\\", \\\\\\\"/org/freedesktop/systemd1/unit/var_2dlib_2dcontainers_2emount\\\\\\\"}, Sequence:0x3d}\\\\nI1206 15:46:34.430033 6139 ovnkube.go:595] Stopping ovnkube...\\\\nI1206 15:46:34.430039 6139 ud\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.820823 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.836032 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.853789 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.867094 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.867144 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.867163 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.867197 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.867215 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:35Z","lastTransitionTime":"2025-12-06T15:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.881886 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.881932 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.881952 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.881977 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.881995 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:35Z","lastTransitionTime":"2025-12-06T15:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.883574 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.905724 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a379c66da38cd6d16eb0fdc5941e3fd5d4127b7cce84fe2bde55918ef2c99c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: E1206 15:46:35.905531 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.913140 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.913172 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.913185 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.913201 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.913214 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:35Z","lastTransitionTime":"2025-12-06T15:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.930778 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f228acd2a1febab7d6991054e11cf509ff83ab26e537db82ab49688144a4a233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: E1206 15:46:35.947687 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.948063 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.953634 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.953816 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.953877 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.953955 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.954014 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:35Z","lastTransitionTime":"2025-12-06T15:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:35 crc kubenswrapper[4813]: E1206 15:46:35.966118 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.966768 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.971087 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.971120 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.971131 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.971148 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.971160 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:35Z","lastTransitionTime":"2025-12-06T15:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.976562 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ab1d9f15bba233c4bbd04ece144a5a8119c168abaeee7d4544269b55ad58ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b14a8cf06c49b6922327f25905466f86e1a0fd0a0b852f751bac6f8cd7db9e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bd4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: E1206 15:46:35.981499 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.987213 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.987248 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.987280 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.987298 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.987308 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:35Z","lastTransitionTime":"2025-12-06T15:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:35 crc kubenswrapper[4813]: I1206 15:46:35.987809 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kzbhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18a94103-5ed7-4e27-a58d-7c989b0f70f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kzbhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: E1206 15:46:35.999442 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:35Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:35 crc kubenswrapper[4813]: E1206 15:46:35.999552 4813 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.000844 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.000878 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.000891 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.000907 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.000917 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:36Z","lastTransitionTime":"2025-12-06T15:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.103367 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.103416 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.103433 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.103455 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.103471 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:36Z","lastTransitionTime":"2025-12-06T15:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.169209 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18a94103-5ed7-4e27-a58d-7c989b0f70f3-metrics-certs\") pod \"network-metrics-daemon-kzbhb\" (UID: \"18a94103-5ed7-4e27-a58d-7c989b0f70f3\") " pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:46:36 crc kubenswrapper[4813]: E1206 15:46:36.169402 4813 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 15:46:36 crc kubenswrapper[4813]: E1206 15:46:36.169473 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18a94103-5ed7-4e27-a58d-7c989b0f70f3-metrics-certs podName:18a94103-5ed7-4e27-a58d-7c989b0f70f3 nodeName:}" failed. No retries permitted until 2025-12-06 15:46:37.169451906 +0000 UTC m=+37.060331512 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/18a94103-5ed7-4e27-a58d-7c989b0f70f3-metrics-certs") pod "network-metrics-daemon-kzbhb" (UID: "18a94103-5ed7-4e27-a58d-7c989b0f70f3") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.206447 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.206503 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.206520 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.206544 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.206561 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:36Z","lastTransitionTime":"2025-12-06T15:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.310075 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.310133 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.310149 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.310175 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.310195 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:36Z","lastTransitionTime":"2025-12-06T15:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.378668 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xdt4d_0e54e31f-e9f8-4dd5-8738-c785df002c93/ovnkube-controller/1.log" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.391308 4813 scope.go:117] "RemoveContainer" containerID="ccdd42da55ee2466605cb3f7dd612f6da8e15ec9f5305f7dfd06d0928aad263f" Dec 06 15:46:36 crc kubenswrapper[4813]: E1206 15:46:36.391544 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-xdt4d_openshift-ovn-kubernetes(0e54e31f-e9f8-4dd5-8738-c785df002c93)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.413785 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.413871 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.413895 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.413921 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.413939 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:36Z","lastTransitionTime":"2025-12-06T15:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.423939 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:36Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.439316 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ab1d9f15bba233c4bbd04ece144a5a8119c168abaeee7d4544269b55ad58ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b14a8cf06c49b6922327f25905466f86e1a0fd0a0b852f751bac6f8cd7db9e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bd4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:36Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.456307 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kzbhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18a94103-5ed7-4e27-a58d-7c989b0f70f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kzbhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:36Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.478550 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:36Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.500359 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:36Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.517054 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.517122 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.517145 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.517176 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.517199 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:36Z","lastTransitionTime":"2025-12-06T15:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.523156 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:36Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.543847 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:36Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.560049 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:36Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.592855 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdd42da55ee2466605cb3f7dd612f6da8e15ec9f5305f7dfd06d0928aad263f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccdd42da55ee2466605cb3f7dd612f6da8e15ec9f5305f7dfd06d0928aad263f\\\",\\\"exitCode\\\":2,\\\"finishedAt\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"message\\\":\\\"/127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z\\\\nI1206 15:46:34.429939 6139 nad_controller.go:166] [node-nad-controller NAD controller]: shutting down\\\\nI1206 15:46:34.429795 6139 lb_config.go:1031] Cluster endpoints for openshift-kube-storage-version-migrator-operator/metrics for network=default are: map[]\\\\nI1206 15:46:34.429993 6139 services_controller.go:443] Built service openshift-kube-storage-version-migrator-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.36\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1206 15:46:34.429986 6139 udn_isolation.go:361] D-Bus event received: \\\\u0026dbus.Signal{Sender:\\\\\\\"org.freedesktop.systemd1\\\\\\\", Path:\\\\\\\"/org/freedesktop/systemd1\\\\\\\", Name:\\\\\\\"org.freedesktop.systemd1.Manager.UnitNew\\\\\\\", Body:[]interface {}{\\\\\\\"var-lib-containers.mount\\\\\\\", \\\\\\\"/org/freedesktop/systemd1/unit/var_2dlib_2dcontainers_2emount\\\\\\\"}, Sequence:0x3d}\\\\nI1206 15:46:34.430033 6139 ovnkube.go:595] Stopping ovnkube...\\\\nI1206 15:46:34.430039 6139 ud\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-xdt4d_openshift-ovn-kubernetes(0e54e31f-e9f8-4dd5-8738-c785df002c93)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:36Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.608960 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:36Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.622515 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.622554 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.622589 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.622611 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.622632 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:36Z","lastTransitionTime":"2025-12-06T15:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.630241 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:36Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.644711 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:36Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.665900 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:36Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.684085 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:36Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.699028 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a379c66da38cd6d16eb0fdc5941e3fd5d4127b7cce84fe2bde55918ef2c99c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:36Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.719325 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f228acd2a1febab7d6991054e11cf509ff83ab26e537db82ab49688144a4a233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:36Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.724898 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.725066 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.725155 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.725281 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.725430 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:36Z","lastTransitionTime":"2025-12-06T15:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.737084 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:36Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.828362 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.828414 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.828431 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.828454 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.828471 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:36Z","lastTransitionTime":"2025-12-06T15:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.931213 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.931475 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.931553 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.931630 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:36 crc kubenswrapper[4813]: I1206 15:46:36.931713 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:36Z","lastTransitionTime":"2025-12-06T15:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.034957 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.035025 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.035043 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.035069 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.035086 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:37Z","lastTransitionTime":"2025-12-06T15:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.138009 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.138087 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.138106 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.138135 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.138158 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:37Z","lastTransitionTime":"2025-12-06T15:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.178835 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18a94103-5ed7-4e27-a58d-7c989b0f70f3-metrics-certs\") pod \"network-metrics-daemon-kzbhb\" (UID: \"18a94103-5ed7-4e27-a58d-7c989b0f70f3\") " pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:46:37 crc kubenswrapper[4813]: E1206 15:46:37.179032 4813 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 15:46:37 crc kubenswrapper[4813]: E1206 15:46:37.179142 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18a94103-5ed7-4e27-a58d-7c989b0f70f3-metrics-certs podName:18a94103-5ed7-4e27-a58d-7c989b0f70f3 nodeName:}" failed. No retries permitted until 2025-12-06 15:46:39.179115109 +0000 UTC m=+39.069994715 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/18a94103-5ed7-4e27-a58d-7c989b0f70f3-metrics-certs") pod "network-metrics-daemon-kzbhb" (UID: "18a94103-5ed7-4e27-a58d-7c989b0f70f3") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.242006 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.242669 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.242822 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.243027 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.243318 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:37Z","lastTransitionTime":"2025-12-06T15:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.346345 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.346416 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.346435 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.346460 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.346478 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:37Z","lastTransitionTime":"2025-12-06T15:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.449978 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.450025 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.450041 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.450062 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.450074 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:37Z","lastTransitionTime":"2025-12-06T15:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.485830 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.485834 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:46:37 crc kubenswrapper[4813]: E1206 15:46:37.485955 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.485836 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:46:37 crc kubenswrapper[4813]: E1206 15:46:37.486082 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.486102 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:46:37 crc kubenswrapper[4813]: E1206 15:46:37.486325 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:46:37 crc kubenswrapper[4813]: E1206 15:46:37.486414 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.552770 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.553102 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.553250 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.553446 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.553626 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:37Z","lastTransitionTime":"2025-12-06T15:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.656855 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.656928 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.656948 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.656973 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.656991 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:37Z","lastTransitionTime":"2025-12-06T15:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.761101 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.761166 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.761185 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.761213 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.761230 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:37Z","lastTransitionTime":"2025-12-06T15:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.864481 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.864580 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.864598 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.864623 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.864642 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:37Z","lastTransitionTime":"2025-12-06T15:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.967785 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.967833 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.967852 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.967876 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:37 crc kubenswrapper[4813]: I1206 15:46:37.967894 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:37Z","lastTransitionTime":"2025-12-06T15:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.071355 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.071415 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.071433 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.071455 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.071476 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:38Z","lastTransitionTime":"2025-12-06T15:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.174810 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.174880 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.174907 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.174940 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.174962 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:38Z","lastTransitionTime":"2025-12-06T15:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.277742 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.277805 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.277821 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.277845 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.277870 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:38Z","lastTransitionTime":"2025-12-06T15:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.381580 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.381639 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.381656 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.381682 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.381705 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:38Z","lastTransitionTime":"2025-12-06T15:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.484873 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.484972 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.485015 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.485043 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.485064 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:38Z","lastTransitionTime":"2025-12-06T15:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.588420 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.588497 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.588516 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.588545 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.588571 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:38Z","lastTransitionTime":"2025-12-06T15:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.692529 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.692581 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.692593 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.692613 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.692628 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:38Z","lastTransitionTime":"2025-12-06T15:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.797795 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.797875 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.797901 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.797934 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.797968 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:38Z","lastTransitionTime":"2025-12-06T15:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.901658 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.901724 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.901740 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.901767 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:38 crc kubenswrapper[4813]: I1206 15:46:38.901784 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:38Z","lastTransitionTime":"2025-12-06T15:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.004000 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.004241 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.004360 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.004483 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.004572 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:39Z","lastTransitionTime":"2025-12-06T15:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.108441 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.108524 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.108549 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.108580 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.108604 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:39Z","lastTransitionTime":"2025-12-06T15:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.199446 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18a94103-5ed7-4e27-a58d-7c989b0f70f3-metrics-certs\") pod \"network-metrics-daemon-kzbhb\" (UID: \"18a94103-5ed7-4e27-a58d-7c989b0f70f3\") " pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:46:39 crc kubenswrapper[4813]: E1206 15:46:39.199699 4813 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 15:46:39 crc kubenswrapper[4813]: E1206 15:46:39.199851 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18a94103-5ed7-4e27-a58d-7c989b0f70f3-metrics-certs podName:18a94103-5ed7-4e27-a58d-7c989b0f70f3 nodeName:}" failed. No retries permitted until 2025-12-06 15:46:43.199817569 +0000 UTC m=+43.090697175 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/18a94103-5ed7-4e27-a58d-7c989b0f70f3-metrics-certs") pod "network-metrics-daemon-kzbhb" (UID: "18a94103-5ed7-4e27-a58d-7c989b0f70f3") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.211835 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.211889 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.211906 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.211932 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.211950 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:39Z","lastTransitionTime":"2025-12-06T15:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.314622 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.314908 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.314998 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.315129 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.315230 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:39Z","lastTransitionTime":"2025-12-06T15:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.418457 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.418551 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.418568 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.418590 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.418608 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:39Z","lastTransitionTime":"2025-12-06T15:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.486698 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.486732 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.486768 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.486831 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:39 crc kubenswrapper[4813]: E1206 15:46:39.486850 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:46:39 crc kubenswrapper[4813]: E1206 15:46:39.487013 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:46:39 crc kubenswrapper[4813]: E1206 15:46:39.487106 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:46:39 crc kubenswrapper[4813]: E1206 15:46:39.487244 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.521357 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.521427 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.521441 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.521459 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.521471 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:39Z","lastTransitionTime":"2025-12-06T15:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.624082 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.624140 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.624153 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.624173 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.624195 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:39Z","lastTransitionTime":"2025-12-06T15:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.727676 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.727745 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.727770 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.727797 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.727819 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:39Z","lastTransitionTime":"2025-12-06T15:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.830887 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.830925 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.830969 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.830987 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.831026 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:39Z","lastTransitionTime":"2025-12-06T15:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.934631 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.934693 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.934710 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.934736 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:39 crc kubenswrapper[4813]: I1206 15:46:39.934755 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:39Z","lastTransitionTime":"2025-12-06T15:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.037980 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.038076 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.038096 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.038123 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.038146 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:40Z","lastTransitionTime":"2025-12-06T15:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.141544 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.141619 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.141637 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.141664 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.141681 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:40Z","lastTransitionTime":"2025-12-06T15:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.247318 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.247417 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.247439 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.247469 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.247533 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:40Z","lastTransitionTime":"2025-12-06T15:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.350703 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.351067 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.351317 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.351561 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.351903 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:40Z","lastTransitionTime":"2025-12-06T15:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.455878 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.455954 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.455979 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.456010 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.456033 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:40Z","lastTransitionTime":"2025-12-06T15:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.514240 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.536969 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.554854 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.559403 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.559458 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.559471 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.559488 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.559520 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:40Z","lastTransitionTime":"2025-12-06T15:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.585010 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a379c66da38cd6d16eb0fdc5941e3fd5d4127b7cce84fe2bde55918ef2c99c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.606591 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f228acd2a1febab7d6991054e11cf509ff83ab26e537db82ab49688144a4a233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.633451 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.655919 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ab1d9f15bba233c4bbd04ece144a5a8119c168abaeee7d4544269b55ad58ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b14a8cf06c49b6922327f25905466f86e1a0fd0a0b852f751bac6f8cd7db9e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bd4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.662384 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.662503 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.662527 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.662546 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.662587 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:40Z","lastTransitionTime":"2025-12-06T15:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.675426 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kzbhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18a94103-5ed7-4e27-a58d-7c989b0f70f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kzbhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.691874 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.725110 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdd42da55ee2466605cb3f7dd612f6da8e15ec9f5305f7dfd06d0928aad263f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccdd42da55ee2466605cb3f7dd612f6da8e15ec9f5305f7dfd06d0928aad263f\\\",\\\"exitCode\\\":2,\\\"finishedAt\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"message\\\":\\\"/127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z\\\\nI1206 15:46:34.429939 6139 nad_controller.go:166] [node-nad-controller NAD controller]: shutting down\\\\nI1206 15:46:34.429795 6139 lb_config.go:1031] Cluster endpoints for openshift-kube-storage-version-migrator-operator/metrics for network=default are: map[]\\\\nI1206 15:46:34.429993 6139 services_controller.go:443] Built service openshift-kube-storage-version-migrator-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.36\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1206 15:46:34.429986 6139 udn_isolation.go:361] D-Bus event received: \\\\u0026dbus.Signal{Sender:\\\\\\\"org.freedesktop.systemd1\\\\\\\", Path:\\\\\\\"/org/freedesktop/systemd1\\\\\\\", Name:\\\\\\\"org.freedesktop.systemd1.Manager.UnitNew\\\\\\\", Body:[]interface {}{\\\\\\\"var-lib-containers.mount\\\\\\\", \\\\\\\"/org/freedesktop/systemd1/unit/var_2dlib_2dcontainers_2emount\\\\\\\"}, Sequence:0x3d}\\\\nI1206 15:46:34.430033 6139 ovnkube.go:595] Stopping ovnkube...\\\\nI1206 15:46:34.430039 6139 ud\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-xdt4d_openshift-ovn-kubernetes(0e54e31f-e9f8-4dd5-8738-c785df002c93)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.743512 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.760338 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.765208 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.765247 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.765455 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.765502 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.765520 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:40Z","lastTransitionTime":"2025-12-06T15:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.775859 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.789685 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.803162 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.817367 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.832314 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.868285 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.868352 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.868367 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.868388 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.868403 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:40Z","lastTransitionTime":"2025-12-06T15:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.971806 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.971875 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.971897 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.971922 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:40 crc kubenswrapper[4813]: I1206 15:46:40.971940 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:40Z","lastTransitionTime":"2025-12-06T15:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.076071 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.076145 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.076166 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.076193 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.076214 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:41Z","lastTransitionTime":"2025-12-06T15:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.179986 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.180061 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.180079 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.180106 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.180127 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:41Z","lastTransitionTime":"2025-12-06T15:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.284358 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.284449 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.284473 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.284505 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.284528 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:41Z","lastTransitionTime":"2025-12-06T15:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.387451 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.387520 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.387582 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.387674 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.387738 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:41Z","lastTransitionTime":"2025-12-06T15:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.486597 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.486729 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.486618 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:46:41 crc kubenswrapper[4813]: E1206 15:46:41.486827 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.486619 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:46:41 crc kubenswrapper[4813]: E1206 15:46:41.486935 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:46:41 crc kubenswrapper[4813]: E1206 15:46:41.487165 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:46:41 crc kubenswrapper[4813]: E1206 15:46:41.487368 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.491869 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.491926 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.491949 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.491974 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.491995 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:41Z","lastTransitionTime":"2025-12-06T15:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.595684 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.596653 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.596820 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.596948 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.597115 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:41Z","lastTransitionTime":"2025-12-06T15:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.700245 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.700352 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.700378 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.700403 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.700419 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:41Z","lastTransitionTime":"2025-12-06T15:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.803774 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.803844 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.803867 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.803895 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.803917 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:41Z","lastTransitionTime":"2025-12-06T15:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.907441 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.907526 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.907555 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.907588 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:41 crc kubenswrapper[4813]: I1206 15:46:41.907615 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:41Z","lastTransitionTime":"2025-12-06T15:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.010630 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.010698 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.010719 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.010744 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.010761 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:42Z","lastTransitionTime":"2025-12-06T15:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.114543 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.114620 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.114643 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.114666 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.114683 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:42Z","lastTransitionTime":"2025-12-06T15:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.217221 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.217305 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.217322 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.217343 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.217364 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:42Z","lastTransitionTime":"2025-12-06T15:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.320341 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.320396 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.320414 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.320436 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.320456 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:42Z","lastTransitionTime":"2025-12-06T15:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.423352 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.423415 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.423432 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.423456 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.423495 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:42Z","lastTransitionTime":"2025-12-06T15:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.526394 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.526459 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.526481 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.526502 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.526518 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:42Z","lastTransitionTime":"2025-12-06T15:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.630503 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.630563 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.630585 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.630612 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.630634 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:42Z","lastTransitionTime":"2025-12-06T15:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.733918 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.733998 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.734021 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.734047 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.734068 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:42Z","lastTransitionTime":"2025-12-06T15:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.837221 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.837651 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.837803 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.837943 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.838070 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:42Z","lastTransitionTime":"2025-12-06T15:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.941785 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.941865 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.941889 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.942440 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:42 crc kubenswrapper[4813]: I1206 15:46:42.942469 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:42Z","lastTransitionTime":"2025-12-06T15:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.047949 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.048000 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.048017 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.048040 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.048056 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:43Z","lastTransitionTime":"2025-12-06T15:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.152003 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.152077 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.152097 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.152125 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.152145 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:43Z","lastTransitionTime":"2025-12-06T15:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.249204 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18a94103-5ed7-4e27-a58d-7c989b0f70f3-metrics-certs\") pod \"network-metrics-daemon-kzbhb\" (UID: \"18a94103-5ed7-4e27-a58d-7c989b0f70f3\") " pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:46:43 crc kubenswrapper[4813]: E1206 15:46:43.249447 4813 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 15:46:43 crc kubenswrapper[4813]: E1206 15:46:43.249534 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18a94103-5ed7-4e27-a58d-7c989b0f70f3-metrics-certs podName:18a94103-5ed7-4e27-a58d-7c989b0f70f3 nodeName:}" failed. No retries permitted until 2025-12-06 15:46:51.249510652 +0000 UTC m=+51.140390258 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/18a94103-5ed7-4e27-a58d-7c989b0f70f3-metrics-certs") pod "network-metrics-daemon-kzbhb" (UID: "18a94103-5ed7-4e27-a58d-7c989b0f70f3") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.255533 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.255648 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.255677 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.255711 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.255733 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:43Z","lastTransitionTime":"2025-12-06T15:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.359214 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.359323 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.359377 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.359412 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.359435 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:43Z","lastTransitionTime":"2025-12-06T15:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.461544 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.461616 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.461641 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.461675 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.461701 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:43Z","lastTransitionTime":"2025-12-06T15:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.485713 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:46:43 crc kubenswrapper[4813]: E1206 15:46:43.486032 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.485865 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:43 crc kubenswrapper[4813]: E1206 15:46:43.486208 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.485814 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:46:43 crc kubenswrapper[4813]: E1206 15:46:43.486424 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.485883 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:46:43 crc kubenswrapper[4813]: E1206 15:46:43.486586 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.564078 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.564141 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.564153 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.564171 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.564202 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:43Z","lastTransitionTime":"2025-12-06T15:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.666623 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.666687 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.666710 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.666740 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.666760 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:43Z","lastTransitionTime":"2025-12-06T15:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.769030 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.769064 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.769072 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.769087 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.769096 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:43Z","lastTransitionTime":"2025-12-06T15:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.871253 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.871309 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.871318 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.871331 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.871358 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:43Z","lastTransitionTime":"2025-12-06T15:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.974242 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.974344 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.974361 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.974390 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:43 crc kubenswrapper[4813]: I1206 15:46:43.974408 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:43Z","lastTransitionTime":"2025-12-06T15:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.078707 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.078791 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.078817 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.078844 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.078862 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:44Z","lastTransitionTime":"2025-12-06T15:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.181744 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.182114 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.182288 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.182485 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.182696 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:44Z","lastTransitionTime":"2025-12-06T15:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.285995 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.286065 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.286086 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.286112 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.286129 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:44Z","lastTransitionTime":"2025-12-06T15:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.388582 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.388627 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.388645 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.388668 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.388684 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:44Z","lastTransitionTime":"2025-12-06T15:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.492585 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.492649 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.492668 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.492692 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.492715 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:44Z","lastTransitionTime":"2025-12-06T15:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.596219 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.596304 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.596329 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.596351 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.596368 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:44Z","lastTransitionTime":"2025-12-06T15:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.699491 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.699549 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.699567 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.699591 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.699608 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:44Z","lastTransitionTime":"2025-12-06T15:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.802645 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.803002 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.803202 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.803437 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.803596 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:44Z","lastTransitionTime":"2025-12-06T15:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.907211 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.907302 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.907320 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.907346 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:44 crc kubenswrapper[4813]: I1206 15:46:44.907365 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:44Z","lastTransitionTime":"2025-12-06T15:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.010719 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.010775 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.010799 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.010829 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.010852 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:45Z","lastTransitionTime":"2025-12-06T15:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.114093 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.114131 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.114142 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.114158 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.114169 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:45Z","lastTransitionTime":"2025-12-06T15:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.217464 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.217518 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.217534 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.217556 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.217574 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:45Z","lastTransitionTime":"2025-12-06T15:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.320749 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.321098 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.321474 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.321826 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.322170 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:45Z","lastTransitionTime":"2025-12-06T15:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.425322 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.425603 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.425743 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.425877 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.425993 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:45Z","lastTransitionTime":"2025-12-06T15:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.486673 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:46:45 crc kubenswrapper[4813]: E1206 15:46:45.486827 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.487194 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.486698 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.487305 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:45 crc kubenswrapper[4813]: E1206 15:46:45.487829 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:46:45 crc kubenswrapper[4813]: E1206 15:46:45.487895 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:46:45 crc kubenswrapper[4813]: E1206 15:46:45.487961 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.529033 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.529097 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.529110 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.529126 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.529138 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:45Z","lastTransitionTime":"2025-12-06T15:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.632112 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.632453 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.632543 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.632639 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.632754 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:45Z","lastTransitionTime":"2025-12-06T15:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.735561 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.735620 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.735637 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.735660 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.735682 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:45Z","lastTransitionTime":"2025-12-06T15:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.839134 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.839197 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.839213 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.839237 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.839254 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:45Z","lastTransitionTime":"2025-12-06T15:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.942299 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.942679 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.942814 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.942951 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:45 crc kubenswrapper[4813]: I1206 15:46:45.943074 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:45Z","lastTransitionTime":"2025-12-06T15:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.045880 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.046129 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.046230 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.046338 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.046418 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:46Z","lastTransitionTime":"2025-12-06T15:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.059382 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.059660 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.059686 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.059710 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.059728 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:46Z","lastTransitionTime":"2025-12-06T15:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:46 crc kubenswrapper[4813]: E1206 15:46:46.075759 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:46Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.080209 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.080249 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.080298 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.080323 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.080340 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:46Z","lastTransitionTime":"2025-12-06T15:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:46 crc kubenswrapper[4813]: E1206 15:46:46.107127 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:46Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.111842 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.111896 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.111914 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.111938 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.111955 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:46Z","lastTransitionTime":"2025-12-06T15:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:46 crc kubenswrapper[4813]: E1206 15:46:46.133409 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:46Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.138578 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.138642 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.138654 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.138677 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.138689 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:46Z","lastTransitionTime":"2025-12-06T15:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:46 crc kubenswrapper[4813]: E1206 15:46:46.158084 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:46Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.163342 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.163492 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.163587 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.163671 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.163746 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:46Z","lastTransitionTime":"2025-12-06T15:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:46 crc kubenswrapper[4813]: E1206 15:46:46.183134 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:46Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:46 crc kubenswrapper[4813]: E1206 15:46:46.183427 4813 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.185159 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.185315 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.185399 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.185499 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.185587 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:46Z","lastTransitionTime":"2025-12-06T15:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.288075 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.288151 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.288165 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.288184 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.288198 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:46Z","lastTransitionTime":"2025-12-06T15:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.390960 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.391030 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.391056 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.391087 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.391107 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:46Z","lastTransitionTime":"2025-12-06T15:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.492596 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.492660 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.492672 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.492687 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.492700 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:46Z","lastTransitionTime":"2025-12-06T15:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.595461 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.595492 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.595501 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.595513 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.595541 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:46Z","lastTransitionTime":"2025-12-06T15:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.698444 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.698486 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.698499 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.698515 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.698537 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:46Z","lastTransitionTime":"2025-12-06T15:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.802483 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.803420 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.803469 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.803530 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.803550 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:46Z","lastTransitionTime":"2025-12-06T15:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.906870 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.906925 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.906942 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.906961 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:46 crc kubenswrapper[4813]: I1206 15:46:46.906976 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:46Z","lastTransitionTime":"2025-12-06T15:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.010025 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.010079 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.010096 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.010118 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.010132 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:47Z","lastTransitionTime":"2025-12-06T15:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.113696 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.113790 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.113817 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.113848 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.113870 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:47Z","lastTransitionTime":"2025-12-06T15:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.216769 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.216842 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.216860 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.216884 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.216901 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:47Z","lastTransitionTime":"2025-12-06T15:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.319980 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.320045 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.320069 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.320099 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.320122 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:47Z","lastTransitionTime":"2025-12-06T15:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.423901 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.423960 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.423979 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.424007 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.424031 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:47Z","lastTransitionTime":"2025-12-06T15:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.486677 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.486787 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:46:47 crc kubenswrapper[4813]: E1206 15:46:47.486884 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.486928 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.486978 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:47 crc kubenswrapper[4813]: E1206 15:46:47.487060 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:46:47 crc kubenswrapper[4813]: E1206 15:46:47.487139 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:46:47 crc kubenswrapper[4813]: E1206 15:46:47.487761 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.488456 4813 scope.go:117] "RemoveContainer" containerID="ccdd42da55ee2466605cb3f7dd612f6da8e15ec9f5305f7dfd06d0928aad263f" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.527094 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.527151 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.527171 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.527195 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.527211 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:47Z","lastTransitionTime":"2025-12-06T15:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.631085 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.631332 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.631343 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.631362 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.631373 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:47Z","lastTransitionTime":"2025-12-06T15:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.734864 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.734909 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.734926 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.734944 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.734963 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:47Z","lastTransitionTime":"2025-12-06T15:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.837912 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.837958 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.837970 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.837986 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.837998 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:47Z","lastTransitionTime":"2025-12-06T15:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.942210 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.942329 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.942356 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.942383 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:47 crc kubenswrapper[4813]: I1206 15:46:47.942416 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:47Z","lastTransitionTime":"2025-12-06T15:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.046237 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.046337 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.046360 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.046404 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.046427 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:48Z","lastTransitionTime":"2025-12-06T15:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.149725 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.149792 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.149812 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.149841 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.149861 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:48Z","lastTransitionTime":"2025-12-06T15:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.293926 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.293998 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.294024 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.294054 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.294073 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:48Z","lastTransitionTime":"2025-12-06T15:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.396817 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.396848 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.396858 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.396875 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.396885 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:48Z","lastTransitionTime":"2025-12-06T15:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.432564 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xdt4d_0e54e31f-e9f8-4dd5-8738-c785df002c93/ovnkube-controller/1.log" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.434231 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" event={"ID":"0e54e31f-e9f8-4dd5-8738-c785df002c93","Type":"ContainerStarted","Data":"0d6b3583626fb73a7ee10d9ea994c5dc388215022154a5a51777e14a8195cc0c"} Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.435069 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.454181 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:48Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.467369 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ab1d9f15bba233c4bbd04ece144a5a8119c168abaeee7d4544269b55ad58ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b14a8cf06c49b6922327f25905466f86e1a0fd0a0b852f751bac6f8cd7db9e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bd4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:48Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.481957 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kzbhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18a94103-5ed7-4e27-a58d-7c989b0f70f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kzbhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:48Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.498913 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.498948 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.498959 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.498972 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.498983 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:48Z","lastTransitionTime":"2025-12-06T15:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.506421 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:48Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.522510 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:48Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.538567 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:48Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.550604 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:48Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.574483 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:48Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.656611 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d6b3583626fb73a7ee10d9ea994c5dc388215022154a5a51777e14a8195cc0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccdd42da55ee2466605cb3f7dd612f6da8e15ec9f5305f7dfd06d0928aad263f\\\",\\\"exitCode\\\":2,\\\"finishedAt\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"message\\\":\\\"/127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z\\\\nI1206 15:46:34.429939 6139 nad_controller.go:166] [node-nad-controller NAD controller]: shutting down\\\\nI1206 15:46:34.429795 6139 lb_config.go:1031] Cluster endpoints for openshift-kube-storage-version-migrator-operator/metrics for network=default are: map[]\\\\nI1206 15:46:34.429993 6139 services_controller.go:443] Built service openshift-kube-storage-version-migrator-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.36\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1206 15:46:34.429986 6139 udn_isolation.go:361] D-Bus event received: \\\\u0026dbus.Signal{Sender:\\\\\\\"org.freedesktop.systemd1\\\\\\\", Path:\\\\\\\"/org/freedesktop/systemd1\\\\\\\", Name:\\\\\\\"org.freedesktop.systemd1.Manager.UnitNew\\\\\\\", Body:[]interface {}{\\\\\\\"var-lib-containers.mount\\\\\\\", \\\\\\\"/org/freedesktop/systemd1/unit/var_2dlib_2dcontainers_2emount\\\\\\\"}, Sequence:0x3d}\\\\nI1206 15:46:34.430033 6139 ovnkube.go:595] Stopping ovnkube...\\\\nI1206 15:46:34.430039 6139 ud\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:48Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.657816 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.657850 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.657862 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.657878 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.657890 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:48Z","lastTransitionTime":"2025-12-06T15:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.668566 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:48Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.680162 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:48Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.694619 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:48Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.710291 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:48Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.725797 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:48Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.744149 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a379c66da38cd6d16eb0fdc5941e3fd5d4127b7cce84fe2bde55918ef2c99c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:48Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.762910 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.762950 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.762961 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.762978 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.762989 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:48Z","lastTransitionTime":"2025-12-06T15:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.812807 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f228acd2a1febab7d6991054e11cf509ff83ab26e537db82ab49688144a4a233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:48Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.828293 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:48Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.865030 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.865062 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.865069 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.865085 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.865095 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:48Z","lastTransitionTime":"2025-12-06T15:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.966990 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.967017 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.967026 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.967040 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:48 crc kubenswrapper[4813]: I1206 15:46:48.967049 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:48Z","lastTransitionTime":"2025-12-06T15:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.069017 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.069055 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.069064 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.069080 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.069090 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:49Z","lastTransitionTime":"2025-12-06T15:46:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.173138 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.173191 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.173205 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.173225 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.173237 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:49Z","lastTransitionTime":"2025-12-06T15:46:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.276170 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.276202 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.276212 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.276231 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.276243 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:49Z","lastTransitionTime":"2025-12-06T15:46:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.378321 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.378362 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.378372 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.378388 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.378398 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:49Z","lastTransitionTime":"2025-12-06T15:46:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.481320 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.481377 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.481388 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.481406 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.481418 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:49Z","lastTransitionTime":"2025-12-06T15:46:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.486480 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.486533 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.486525 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:46:49 crc kubenswrapper[4813]: E1206 15:46:49.486632 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.486651 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:49 crc kubenswrapper[4813]: E1206 15:46:49.486832 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:46:49 crc kubenswrapper[4813]: E1206 15:46:49.487002 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:46:49 crc kubenswrapper[4813]: E1206 15:46:49.487215 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.586144 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.586191 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.586203 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.586220 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.586282 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:49Z","lastTransitionTime":"2025-12-06T15:46:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.689754 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.689845 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.689873 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.689903 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.689924 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:49Z","lastTransitionTime":"2025-12-06T15:46:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.792118 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.792447 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.792525 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.792587 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.792664 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:49Z","lastTransitionTime":"2025-12-06T15:46:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.895127 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.895649 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.895722 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.895783 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.895840 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:49Z","lastTransitionTime":"2025-12-06T15:46:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.998545 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.998802 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.998901 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.998992 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:49 crc kubenswrapper[4813]: I1206 15:46:49.999084 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:49Z","lastTransitionTime":"2025-12-06T15:46:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.102298 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.102689 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.102842 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.102993 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.103151 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:50Z","lastTransitionTime":"2025-12-06T15:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.209669 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.209733 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.209751 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.209780 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.209796 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:50Z","lastTransitionTime":"2025-12-06T15:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.312737 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.312797 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.312820 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.312847 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.312864 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:50Z","lastTransitionTime":"2025-12-06T15:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.416177 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.416242 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.416281 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.416306 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.416325 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:50Z","lastTransitionTime":"2025-12-06T15:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.446161 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xdt4d_0e54e31f-e9f8-4dd5-8738-c785df002c93/ovnkube-controller/2.log" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.447398 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xdt4d_0e54e31f-e9f8-4dd5-8738-c785df002c93/ovnkube-controller/1.log" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.451879 4813 generic.go:334] "Generic (PLEG): container finished" podID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerID="0d6b3583626fb73a7ee10d9ea994c5dc388215022154a5a51777e14a8195cc0c" exitCode=1 Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.451999 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" event={"ID":"0e54e31f-e9f8-4dd5-8738-c785df002c93","Type":"ContainerDied","Data":"0d6b3583626fb73a7ee10d9ea994c5dc388215022154a5a51777e14a8195cc0c"} Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.452212 4813 scope.go:117] "RemoveContainer" containerID="ccdd42da55ee2466605cb3f7dd612f6da8e15ec9f5305f7dfd06d0928aad263f" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.453580 4813 scope.go:117] "RemoveContainer" containerID="0d6b3583626fb73a7ee10d9ea994c5dc388215022154a5a51777e14a8195cc0c" Dec 06 15:46:50 crc kubenswrapper[4813]: E1206 15:46:50.453945 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xdt4d_openshift-ovn-kubernetes(0e54e31f-e9f8-4dd5-8738-c785df002c93)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.494558 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:50Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.513192 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ab1d9f15bba233c4bbd04ece144a5a8119c168abaeee7d4544269b55ad58ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b14a8cf06c49b6922327f25905466f86e1a0fd0a0b852f751bac6f8cd7db9e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bd4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:50Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.519052 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.519106 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.519123 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.519148 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.519167 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:50Z","lastTransitionTime":"2025-12-06T15:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.532120 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kzbhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18a94103-5ed7-4e27-a58d-7c989b0f70f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kzbhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:50Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.558906 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:50Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.580000 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:50Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.615637 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d6b3583626fb73a7ee10d9ea994c5dc388215022154a5a51777e14a8195cc0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccdd42da55ee2466605cb3f7dd612f6da8e15ec9f5305f7dfd06d0928aad263f\\\",\\\"exitCode\\\":2,\\\"finishedAt\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"message\\\":\\\"/127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z\\\\nI1206 15:46:34.429939 6139 nad_controller.go:166] [node-nad-controller NAD controller]: shutting down\\\\nI1206 15:46:34.429795 6139 lb_config.go:1031] Cluster endpoints for openshift-kube-storage-version-migrator-operator/metrics for network=default are: map[]\\\\nI1206 15:46:34.429993 6139 services_controller.go:443] Built service openshift-kube-storage-version-migrator-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.36\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1206 15:46:34.429986 6139 udn_isolation.go:361] D-Bus event received: \\\\u0026dbus.Signal{Sender:\\\\\\\"org.freedesktop.systemd1\\\\\\\", Path:\\\\\\\"/org/freedesktop/systemd1\\\\\\\", Name:\\\\\\\"org.freedesktop.systemd1.Manager.UnitNew\\\\\\\", Body:[]interface {}{\\\\\\\"var-lib-containers.mount\\\\\\\", \\\\\\\"/org/freedesktop/systemd1/unit/var_2dlib_2dcontainers_2emount\\\\\\\"}, Sequence:0x3d}\\\\nI1206 15:46:34.430033 6139 ovnkube.go:595] Stopping ovnkube...\\\\nI1206 15:46:34.430039 6139 ud\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d6b3583626fb73a7ee10d9ea994c5dc388215022154a5a51777e14a8195cc0c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T15:46:49Z\\\",\\\"message\\\":\\\"ontroller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:49Z is after 2025-08-24T17:21:41Z]\\\\nI1206 15:46:49.421493 6320 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"18746a4d-8a63-458a-b7e3-8fb89ff95fc0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-route-controller-manager/route-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:50Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.624255 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.624324 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.624334 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.624355 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.624366 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:50Z","lastTransitionTime":"2025-12-06T15:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.635167 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:50Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.648970 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.653023 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:50Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.663320 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.677583 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:50Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.695850 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:50Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.710761 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:50Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.726526 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.726564 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.726572 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.726588 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.726598 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:50Z","lastTransitionTime":"2025-12-06T15:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.727491 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:50Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.746918 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f228acd2a1febab7d6991054e11cf509ff83ab26e537db82ab49688144a4a233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:50Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.762678 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:50Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.778001 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:50Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.800395 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:50Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.816449 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a379c66da38cd6d16eb0fdc5941e3fd5d4127b7cce84fe2bde55918ef2c99c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:50Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.829933 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.829973 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.829985 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.830004 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.830015 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:50Z","lastTransitionTime":"2025-12-06T15:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.833787 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"012e3f9a-67f1-45f2-8e57-76928bb4d5c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a77b5cb2e8c50b39c3369f156a7800603e7d8bf3d6a544ad0f3a6814361338f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2f1dd12351d9f1ce3ccc9711361cf346d9aa1c4cb5d971020e242c80e557eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8326b53a1a99f7cbd09d55008e9c1b4087993a25ae4b50fafc70e4c7b99138d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9471d3db940ee579d5b60cf1f7750c2c265a8f8e549aaa7ed60e82982cf80d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9471d3db940ee579d5b60cf1f7750c2c265a8f8e549aaa7ed60e82982cf80d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:50Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.852563 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:50Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.865643 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:50Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.879840 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:50Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.895562 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:50Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.908592 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a379c66da38cd6d16eb0fdc5941e3fd5d4127b7cce84fe2bde55918ef2c99c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:50Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.930893 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f228acd2a1febab7d6991054e11cf509ff83ab26e537db82ab49688144a4a233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:50Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.932452 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.932606 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.932709 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.932859 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.933010 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:50Z","lastTransitionTime":"2025-12-06T15:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:50 crc kubenswrapper[4813]: I1206 15:46:50.952408 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:50Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.013623 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:50Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.035661 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.035886 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.035947 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.036008 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.036065 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:51Z","lastTransitionTime":"2025-12-06T15:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.039626 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ab1d9f15bba233c4bbd04ece144a5a8119c168abaeee7d4544269b55ad58ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b14a8cf06c49b6922327f25905466f86e1a0fd0a0b852f751bac6f8cd7db9e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bd4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:51Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.055633 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kzbhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18a94103-5ed7-4e27-a58d-7c989b0f70f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kzbhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:51Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.066461 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:51Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.077676 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:51Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.091892 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:51Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.104380 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:51Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.116886 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:51Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.128173 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:51Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.138728 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.138889 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.138975 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.139062 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.139135 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:51Z","lastTransitionTime":"2025-12-06T15:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.152868 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d6b3583626fb73a7ee10d9ea994c5dc388215022154a5a51777e14a8195cc0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccdd42da55ee2466605cb3f7dd612f6da8e15ec9f5305f7dfd06d0928aad263f\\\",\\\"exitCode\\\":2,\\\"finishedAt\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"message\\\":\\\"/127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z\\\\nI1206 15:46:34.429939 6139 nad_controller.go:166] [node-nad-controller NAD controller]: shutting down\\\\nI1206 15:46:34.429795 6139 lb_config.go:1031] Cluster endpoints for openshift-kube-storage-version-migrator-operator/metrics for network=default are: map[]\\\\nI1206 15:46:34.429993 6139 services_controller.go:443] Built service openshift-kube-storage-version-migrator-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.36\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1206 15:46:34.429986 6139 udn_isolation.go:361] D-Bus event received: \\\\u0026dbus.Signal{Sender:\\\\\\\"org.freedesktop.systemd1\\\\\\\", Path:\\\\\\\"/org/freedesktop/systemd1\\\\\\\", Name:\\\\\\\"org.freedesktop.systemd1.Manager.UnitNew\\\\\\\", Body:[]interface {}{\\\\\\\"var-lib-containers.mount\\\\\\\", \\\\\\\"/org/freedesktop/systemd1/unit/var_2dlib_2dcontainers_2emount\\\\\\\"}, Sequence:0x3d}\\\\nI1206 15:46:34.430033 6139 ovnkube.go:595] Stopping ovnkube...\\\\nI1206 15:46:34.430039 6139 ud\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d6b3583626fb73a7ee10d9ea994c5dc388215022154a5a51777e14a8195cc0c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T15:46:49Z\\\",\\\"message\\\":\\\"ontroller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:49Z is after 2025-08-24T17:21:41Z]\\\\nI1206 15:46:49.421493 6320 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"18746a4d-8a63-458a-b7e3-8fb89ff95fc0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-route-controller-manager/route-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:51Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.171888 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"012e3f9a-67f1-45f2-8e57-76928bb4d5c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a77b5cb2e8c50b39c3369f156a7800603e7d8bf3d6a544ad0f3a6814361338f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2f1dd12351d9f1ce3ccc9711361cf346d9aa1c4cb5d971020e242c80e557eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8326b53a1a99f7cbd09d55008e9c1b4087993a25ae4b50fafc70e4c7b99138d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9471d3db940ee579d5b60cf1f7750c2c265a8f8e549aaa7ed60e82982cf80d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9471d3db940ee579d5b60cf1f7750c2c265a8f8e549aaa7ed60e82982cf80d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:51Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.186186 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:51Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.200759 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:51Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.215932 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:51Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.231178 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:51Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.242185 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.242423 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.242496 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.242566 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.242630 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:51Z","lastTransitionTime":"2025-12-06T15:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.250884 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a379c66da38cd6d16eb0fdc5941e3fd5d4127b7cce84fe2bde55918ef2c99c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:51Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.266309 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f228acd2a1febab7d6991054e11cf509ff83ab26e537db82ab49688144a4a233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:51Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.285641 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:51Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.317416 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:51Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.321199 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.321487 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18a94103-5ed7-4e27-a58d-7c989b0f70f3-metrics-certs\") pod \"network-metrics-daemon-kzbhb\" (UID: \"18a94103-5ed7-4e27-a58d-7c989b0f70f3\") " pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:46:51 crc kubenswrapper[4813]: E1206 15:46:51.321530 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:47:23.321495553 +0000 UTC m=+83.212375149 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.321895 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.322040 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:51 crc kubenswrapper[4813]: E1206 15:46:51.321643 4813 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.322161 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:46:51 crc kubenswrapper[4813]: E1206 15:46:51.322113 4813 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 15:46:51 crc kubenswrapper[4813]: E1206 15:46:51.322351 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 15:47:23.322339135 +0000 UTC m=+83.213218731 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 15:46:51 crc kubenswrapper[4813]: E1206 15:46:51.322398 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18a94103-5ed7-4e27-a58d-7c989b0f70f3-metrics-certs podName:18a94103-5ed7-4e27-a58d-7c989b0f70f3 nodeName:}" failed. No retries permitted until 2025-12-06 15:47:07.322362835 +0000 UTC m=+67.213242441 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/18a94103-5ed7-4e27-a58d-7c989b0f70f3-metrics-certs") pod "network-metrics-daemon-kzbhb" (UID: "18a94103-5ed7-4e27-a58d-7c989b0f70f3") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 15:46:51 crc kubenswrapper[4813]: E1206 15:46:51.322137 4813 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 15:46:51 crc kubenswrapper[4813]: E1206 15:46:51.322530 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 15:47:23.322510899 +0000 UTC m=+83.213390585 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 15:46:51 crc kubenswrapper[4813]: E1206 15:46:51.322740 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 15:46:51 crc kubenswrapper[4813]: E1206 15:46:51.322891 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 15:46:51 crc kubenswrapper[4813]: E1206 15:46:51.322994 4813 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:46:51 crc kubenswrapper[4813]: E1206 15:46:51.323118 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 15:47:23.323103944 +0000 UTC m=+83.213983540 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.334544 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ab1d9f15bba233c4bbd04ece144a5a8119c168abaeee7d4544269b55ad58ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b14a8cf06c49b6922327f25905466f86e1a0fd0a0b852f751bac6f8cd7db9e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bd4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:51Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.345674 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.345942 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.346032 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.346121 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.346206 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:51Z","lastTransitionTime":"2025-12-06T15:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.349628 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kzbhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18a94103-5ed7-4e27-a58d-7c989b0f70f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kzbhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:51Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.364677 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:51Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.381850 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:51Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.398863 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:51Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.413176 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:51Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.433387 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:51Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.445518 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:51Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.448474 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.448516 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.448528 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.448550 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.448561 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:51Z","lastTransitionTime":"2025-12-06T15:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.459023 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xdt4d_0e54e31f-e9f8-4dd5-8738-c785df002c93/ovnkube-controller/2.log" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.473451 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d6b3583626fb73a7ee10d9ea994c5dc388215022154a5a51777e14a8195cc0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccdd42da55ee2466605cb3f7dd612f6da8e15ec9f5305f7dfd06d0928aad263f\\\",\\\"exitCode\\\":2,\\\"finishedAt\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"message\\\":\\\"/127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z\\\\nI1206 15:46:34.429939 6139 nad_controller.go:166] [node-nad-controller NAD controller]: shutting down\\\\nI1206 15:46:34.429795 6139 lb_config.go:1031] Cluster endpoints for openshift-kube-storage-version-migrator-operator/metrics for network=default are: map[]\\\\nI1206 15:46:34.429993 6139 services_controller.go:443] Built service openshift-kube-storage-version-migrator-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.36\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1206 15:46:34.429986 6139 udn_isolation.go:361] D-Bus event received: \\\\u0026dbus.Signal{Sender:\\\\\\\"org.freedesktop.systemd1\\\\\\\", Path:\\\\\\\"/org/freedesktop/systemd1\\\\\\\", Name:\\\\\\\"org.freedesktop.systemd1.Manager.UnitNew\\\\\\\", Body:[]interface {}{\\\\\\\"var-lib-containers.mount\\\\\\\", \\\\\\\"/org/freedesktop/systemd1/unit/var_2dlib_2dcontainers_2emount\\\\\\\"}, Sequence:0x3d}\\\\nI1206 15:46:34.430033 6139 ovnkube.go:595] Stopping ovnkube...\\\\nI1206 15:46:34.430039 6139 ud\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d6b3583626fb73a7ee10d9ea994c5dc388215022154a5a51777e14a8195cc0c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T15:46:49Z\\\",\\\"message\\\":\\\"ontroller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:49Z is after 2025-08-24T17:21:41Z]\\\\nI1206 15:46:49.421493 6320 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"18746a4d-8a63-458a-b7e3-8fb89ff95fc0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-route-controller-manager/route-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:51Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.486522 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:51 crc kubenswrapper[4813]: E1206 15:46:51.486691 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.486882 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:46:51 crc kubenswrapper[4813]: E1206 15:46:51.487091 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.486527 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:46:51 crc kubenswrapper[4813]: E1206 15:46:51.487474 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.486951 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:46:51 crc kubenswrapper[4813]: E1206 15:46:51.487842 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.551361 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.551398 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.551410 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.551428 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.551441 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:51Z","lastTransitionTime":"2025-12-06T15:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.624412 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:46:51 crc kubenswrapper[4813]: E1206 15:46:51.624663 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 15:46:51 crc kubenswrapper[4813]: E1206 15:46:51.624698 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 15:46:51 crc kubenswrapper[4813]: E1206 15:46:51.624725 4813 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:46:51 crc kubenswrapper[4813]: E1206 15:46:51.624799 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 15:47:23.624776632 +0000 UTC m=+83.515656238 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.654497 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.654560 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.654582 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.654608 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.654665 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:51Z","lastTransitionTime":"2025-12-06T15:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.756890 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.756944 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.756961 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.756986 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.757005 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:51Z","lastTransitionTime":"2025-12-06T15:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.859901 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.859956 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.859972 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.859995 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.860014 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:51Z","lastTransitionTime":"2025-12-06T15:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.963024 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.963939 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.964367 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.964533 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:51 crc kubenswrapper[4813]: I1206 15:46:51.964708 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:51Z","lastTransitionTime":"2025-12-06T15:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.067214 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.067277 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.067289 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.067302 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.067311 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:52Z","lastTransitionTime":"2025-12-06T15:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.170162 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.170505 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.170711 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.170889 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.171500 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:52Z","lastTransitionTime":"2025-12-06T15:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.273329 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.273367 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.273379 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.273394 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.273405 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:52Z","lastTransitionTime":"2025-12-06T15:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.376500 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.376560 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.376576 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.376598 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.376614 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:52Z","lastTransitionTime":"2025-12-06T15:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.479567 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.479633 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.479655 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.479684 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.479705 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:52Z","lastTransitionTime":"2025-12-06T15:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.582189 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.582233 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.582247 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.582285 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.582298 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:52Z","lastTransitionTime":"2025-12-06T15:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.685781 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.686656 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.686899 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.687158 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.687430 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:52Z","lastTransitionTime":"2025-12-06T15:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.791089 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.791158 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.791179 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.791206 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.791225 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:52Z","lastTransitionTime":"2025-12-06T15:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.894441 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.894516 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.894540 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.894570 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.894607 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:52Z","lastTransitionTime":"2025-12-06T15:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.997866 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.997931 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.997956 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.997989 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:52 crc kubenswrapper[4813]: I1206 15:46:52.998036 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:52Z","lastTransitionTime":"2025-12-06T15:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.101772 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.101838 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.101856 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.101881 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.101899 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:53Z","lastTransitionTime":"2025-12-06T15:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.204412 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.204453 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.204464 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.204480 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.204493 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:53Z","lastTransitionTime":"2025-12-06T15:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.308300 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.308355 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.308374 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.308402 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.308421 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:53Z","lastTransitionTime":"2025-12-06T15:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.411685 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.411752 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.411777 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.411810 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.411831 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:53Z","lastTransitionTime":"2025-12-06T15:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.486325 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.486326 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.486371 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.486457 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:46:53 crc kubenswrapper[4813]: E1206 15:46:53.486612 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:46:53 crc kubenswrapper[4813]: E1206 15:46:53.486710 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:46:53 crc kubenswrapper[4813]: E1206 15:46:53.486809 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:46:53 crc kubenswrapper[4813]: E1206 15:46:53.486884 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.514799 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.514852 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.514869 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.514891 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.514908 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:53Z","lastTransitionTime":"2025-12-06T15:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.618079 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.618141 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.618162 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.618195 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.618220 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:53Z","lastTransitionTime":"2025-12-06T15:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.721423 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.721544 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.721569 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.721642 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.721664 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:53Z","lastTransitionTime":"2025-12-06T15:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.824586 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.824627 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.824642 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.824663 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.824681 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:53Z","lastTransitionTime":"2025-12-06T15:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.927550 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.927593 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.927609 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.927630 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:53 crc kubenswrapper[4813]: I1206 15:46:53.927644 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:53Z","lastTransitionTime":"2025-12-06T15:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.030138 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.030245 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.030325 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.030355 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.030442 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:54Z","lastTransitionTime":"2025-12-06T15:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.133425 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.133495 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.133517 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.133544 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.133561 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:54Z","lastTransitionTime":"2025-12-06T15:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.236898 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.236973 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.236990 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.237012 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.237603 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:54Z","lastTransitionTime":"2025-12-06T15:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.340931 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.340978 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.340995 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.341017 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.341037 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:54Z","lastTransitionTime":"2025-12-06T15:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.444232 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.444381 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.444407 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.444438 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.444459 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:54Z","lastTransitionTime":"2025-12-06T15:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.546951 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.547395 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.547547 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.547701 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.547843 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:54Z","lastTransitionTime":"2025-12-06T15:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.651172 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.651248 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.651304 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.651336 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.651377 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:54Z","lastTransitionTime":"2025-12-06T15:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.753882 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.753936 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.753955 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.753977 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.753993 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:54Z","lastTransitionTime":"2025-12-06T15:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.857668 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.857754 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.857774 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.857822 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.857839 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:54Z","lastTransitionTime":"2025-12-06T15:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.960656 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.960714 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.960731 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.960756 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:54 crc kubenswrapper[4813]: I1206 15:46:54.960774 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:54Z","lastTransitionTime":"2025-12-06T15:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.064092 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.064198 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.064217 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.064242 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.064285 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:55Z","lastTransitionTime":"2025-12-06T15:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.167073 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.167108 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.167119 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.167135 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.167145 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:55Z","lastTransitionTime":"2025-12-06T15:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.269371 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.269408 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.269419 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.269435 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.269446 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:55Z","lastTransitionTime":"2025-12-06T15:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.372389 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.372436 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.372453 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.372476 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.372493 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:55Z","lastTransitionTime":"2025-12-06T15:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.474928 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.474987 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.475008 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.475040 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.475057 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:55Z","lastTransitionTime":"2025-12-06T15:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.486370 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.486413 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.486428 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.486451 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:46:55 crc kubenswrapper[4813]: E1206 15:46:55.486529 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:46:55 crc kubenswrapper[4813]: E1206 15:46:55.486708 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:46:55 crc kubenswrapper[4813]: E1206 15:46:55.486806 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:46:55 crc kubenswrapper[4813]: E1206 15:46:55.486891 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.578670 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.578734 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.578752 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.578776 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.578794 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:55Z","lastTransitionTime":"2025-12-06T15:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.681191 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.681288 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.681317 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.681346 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.681366 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:55Z","lastTransitionTime":"2025-12-06T15:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.784665 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.784718 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.784737 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.784761 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.784777 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:55Z","lastTransitionTime":"2025-12-06T15:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.888494 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.888554 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.888574 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.888597 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.888616 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:55Z","lastTransitionTime":"2025-12-06T15:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.990635 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.991046 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.991143 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.991212 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:55 crc kubenswrapper[4813]: I1206 15:46:55.991301 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:55Z","lastTransitionTime":"2025-12-06T15:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.094203 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.094241 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.094252 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.094425 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.094442 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:56Z","lastTransitionTime":"2025-12-06T15:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.197997 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.198056 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.198074 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.198098 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.198115 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:56Z","lastTransitionTime":"2025-12-06T15:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.288591 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.288935 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.289079 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.289216 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.289451 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:56Z","lastTransitionTime":"2025-12-06T15:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:56 crc kubenswrapper[4813]: E1206 15:46:56.311733 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:56Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.316807 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.316845 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.316855 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.316891 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.316905 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:56Z","lastTransitionTime":"2025-12-06T15:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:56 crc kubenswrapper[4813]: E1206 15:46:56.336044 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:56Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.340837 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.340867 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.340898 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.340915 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.340927 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:56Z","lastTransitionTime":"2025-12-06T15:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:56 crc kubenswrapper[4813]: E1206 15:46:56.359578 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:56Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.364010 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.364077 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.364090 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.364107 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.364121 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:56Z","lastTransitionTime":"2025-12-06T15:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:56 crc kubenswrapper[4813]: E1206 15:46:56.380350 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:56Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.384961 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.385235 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.385522 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.385754 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.385961 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:56Z","lastTransitionTime":"2025-12-06T15:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:56 crc kubenswrapper[4813]: E1206 15:46:56.402238 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:56Z is after 2025-08-24T17:21:41Z" Dec 06 15:46:56 crc kubenswrapper[4813]: E1206 15:46:56.402426 4813 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.404406 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.404442 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.404458 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.404479 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.404494 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:56Z","lastTransitionTime":"2025-12-06T15:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.507187 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.507256 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.507313 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.507344 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.507364 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:56Z","lastTransitionTime":"2025-12-06T15:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.610711 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.610767 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.610780 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.610806 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.610818 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:56Z","lastTransitionTime":"2025-12-06T15:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.714685 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.714743 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.714760 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.714785 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.714805 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:56Z","lastTransitionTime":"2025-12-06T15:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.818183 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.818216 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.818227 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.818244 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.818255 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:56Z","lastTransitionTime":"2025-12-06T15:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.921986 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.922043 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.922059 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.922085 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:56 crc kubenswrapper[4813]: I1206 15:46:56.922103 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:56Z","lastTransitionTime":"2025-12-06T15:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.025457 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.025524 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.025543 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.025568 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.025586 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:57Z","lastTransitionTime":"2025-12-06T15:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.128184 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.128244 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.128316 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.128346 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.128366 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:57Z","lastTransitionTime":"2025-12-06T15:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.232165 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.232600 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.232741 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.232883 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.232998 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:57Z","lastTransitionTime":"2025-12-06T15:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.335834 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.335896 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.335920 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.335955 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.335978 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:57Z","lastTransitionTime":"2025-12-06T15:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.439149 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.439198 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.439209 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.439229 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.439243 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:57Z","lastTransitionTime":"2025-12-06T15:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.486409 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.486465 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.486417 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:46:57 crc kubenswrapper[4813]: E1206 15:46:57.486590 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.486416 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:46:57 crc kubenswrapper[4813]: E1206 15:46:57.486790 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:46:57 crc kubenswrapper[4813]: E1206 15:46:57.486888 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:46:57 crc kubenswrapper[4813]: E1206 15:46:57.486964 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.541807 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.541857 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.541869 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.541886 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.541898 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:57Z","lastTransitionTime":"2025-12-06T15:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.645121 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.645568 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.645813 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.646028 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.646246 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:57Z","lastTransitionTime":"2025-12-06T15:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.749493 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.749542 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.749553 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.749571 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.749582 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:57Z","lastTransitionTime":"2025-12-06T15:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.852764 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.852843 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.852867 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.852902 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.852923 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:57Z","lastTransitionTime":"2025-12-06T15:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.955823 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.955907 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.955925 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.955949 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:57 crc kubenswrapper[4813]: I1206 15:46:57.955970 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:57Z","lastTransitionTime":"2025-12-06T15:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.058983 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.059057 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.059081 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.059112 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.059133 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:58Z","lastTransitionTime":"2025-12-06T15:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.162434 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.162501 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.162529 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.162559 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.162581 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:58Z","lastTransitionTime":"2025-12-06T15:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.266185 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.266240 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.266258 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.266320 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.266337 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:58Z","lastTransitionTime":"2025-12-06T15:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.368720 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.368767 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.368784 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.368807 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.368858 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:58Z","lastTransitionTime":"2025-12-06T15:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.472320 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.472363 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.472374 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.472388 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.472401 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:58Z","lastTransitionTime":"2025-12-06T15:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.574809 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.574855 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.574866 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.574881 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.574892 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:58Z","lastTransitionTime":"2025-12-06T15:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.678229 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.678351 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.678373 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.678401 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.678419 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:58Z","lastTransitionTime":"2025-12-06T15:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.780796 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.780840 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.780850 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.780865 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.780876 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:58Z","lastTransitionTime":"2025-12-06T15:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.884224 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.884307 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.884323 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.884346 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.884363 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:58Z","lastTransitionTime":"2025-12-06T15:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.988224 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.988356 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.988382 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.988407 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:58 crc kubenswrapper[4813]: I1206 15:46:58.988424 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:58Z","lastTransitionTime":"2025-12-06T15:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.091782 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.091889 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.091905 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.091926 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.091941 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:59Z","lastTransitionTime":"2025-12-06T15:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.195452 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.195529 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.195552 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.195584 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.195606 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:59Z","lastTransitionTime":"2025-12-06T15:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.299516 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.299575 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.299592 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.299616 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.299633 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:59Z","lastTransitionTime":"2025-12-06T15:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.402740 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.402801 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.402822 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.402851 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.402872 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:59Z","lastTransitionTime":"2025-12-06T15:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.485941 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.486011 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:46:59 crc kubenswrapper[4813]: E1206 15:46:59.486095 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.486333 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:46:59 crc kubenswrapper[4813]: E1206 15:46:59.486336 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:46:59 crc kubenswrapper[4813]: E1206 15:46:59.486397 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.486478 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:46:59 crc kubenswrapper[4813]: E1206 15:46:59.486540 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.504816 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.504844 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.504854 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.504869 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.504879 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:59Z","lastTransitionTime":"2025-12-06T15:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.607742 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.607770 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.607780 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.607792 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.607800 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:59Z","lastTransitionTime":"2025-12-06T15:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.711088 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.711122 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.711133 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.711147 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.711158 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:59Z","lastTransitionTime":"2025-12-06T15:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.813936 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.814007 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.814062 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.814076 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.814088 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:59Z","lastTransitionTime":"2025-12-06T15:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.917170 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.917238 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.917318 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.917355 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:46:59 crc kubenswrapper[4813]: I1206 15:46:59.917379 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:46:59Z","lastTransitionTime":"2025-12-06T15:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.020530 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.020590 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.020608 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.020632 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.020649 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:00Z","lastTransitionTime":"2025-12-06T15:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.122750 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.122810 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.122828 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.122853 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.122870 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:00Z","lastTransitionTime":"2025-12-06T15:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.228157 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.228237 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.228292 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.228319 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.228337 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:00Z","lastTransitionTime":"2025-12-06T15:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.331864 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.331928 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.331946 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.331970 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.331991 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:00Z","lastTransitionTime":"2025-12-06T15:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.435910 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.435971 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.435987 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.436011 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.436028 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:00Z","lastTransitionTime":"2025-12-06T15:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.508543 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"012e3f9a-67f1-45f2-8e57-76928bb4d5c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a77b5cb2e8c50b39c3369f156a7800603e7d8bf3d6a544ad0f3a6814361338f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2f1dd12351d9f1ce3ccc9711361cf346d9aa1c4cb5d971020e242c80e557eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8326b53a1a99f7cbd09d55008e9c1b4087993a25ae4b50fafc70e4c7b99138d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9471d3db940ee579d5b60cf1f7750c2c265a8f8e549aaa7ed60e82982cf80d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9471d3db940ee579d5b60cf1f7750c2c265a8f8e549aaa7ed60e82982cf80d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:00Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.534589 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:00Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.539714 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.539782 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.539805 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.539834 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.539861 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:00Z","lastTransitionTime":"2025-12-06T15:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.552699 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:00Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.575408 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:00Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.597523 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:00Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.614629 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a379c66da38cd6d16eb0fdc5941e3fd5d4127b7cce84fe2bde55918ef2c99c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:00Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.633964 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f228acd2a1febab7d6991054e11cf509ff83ab26e537db82ab49688144a4a233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:00Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.645827 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.645884 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.645900 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.645922 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.645939 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:00Z","lastTransitionTime":"2025-12-06T15:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.653385 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:00Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.690584 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:00Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.710694 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ab1d9f15bba233c4bbd04ece144a5a8119c168abaeee7d4544269b55ad58ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b14a8cf06c49b6922327f25905466f86e1a0fd0a0b852f751bac6f8cd7db9e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bd4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:00Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.726444 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kzbhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18a94103-5ed7-4e27-a58d-7c989b0f70f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kzbhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:00Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.742707 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:00Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.748042 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.748099 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.748114 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.748135 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.748150 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:00Z","lastTransitionTime":"2025-12-06T15:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.763556 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:00Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.786087 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:00Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.808781 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:00Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.823206 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:00Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.837577 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:00Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.851409 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.851507 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.851527 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.851602 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.851621 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:00Z","lastTransitionTime":"2025-12-06T15:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.868061 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d6b3583626fb73a7ee10d9ea994c5dc388215022154a5a51777e14a8195cc0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccdd42da55ee2466605cb3f7dd612f6da8e15ec9f5305f7dfd06d0928aad263f\\\",\\\"exitCode\\\":2,\\\"finishedAt\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"message\\\":\\\"/127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:34Z is after 2025-08-24T17:21:41Z\\\\nI1206 15:46:34.429939 6139 nad_controller.go:166] [node-nad-controller NAD controller]: shutting down\\\\nI1206 15:46:34.429795 6139 lb_config.go:1031] Cluster endpoints for openshift-kube-storage-version-migrator-operator/metrics for network=default are: map[]\\\\nI1206 15:46:34.429993 6139 services_controller.go:443] Built service openshift-kube-storage-version-migrator-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.36\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1206 15:46:34.429986 6139 udn_isolation.go:361] D-Bus event received: \\\\u0026dbus.Signal{Sender:\\\\\\\"org.freedesktop.systemd1\\\\\\\", Path:\\\\\\\"/org/freedesktop/systemd1\\\\\\\", Name:\\\\\\\"org.freedesktop.systemd1.Manager.UnitNew\\\\\\\", Body:[]interface {}{\\\\\\\"var-lib-containers.mount\\\\\\\", \\\\\\\"/org/freedesktop/systemd1/unit/var_2dlib_2dcontainers_2emount\\\\\\\"}, Sequence:0x3d}\\\\nI1206 15:46:34.430033 6139 ovnkube.go:595] Stopping ovnkube...\\\\nI1206 15:46:34.430039 6139 ud\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d6b3583626fb73a7ee10d9ea994c5dc388215022154a5a51777e14a8195cc0c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T15:46:49Z\\\",\\\"message\\\":\\\"ontroller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:49Z is after 2025-08-24T17:21:41Z]\\\\nI1206 15:46:49.421493 6320 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"18746a4d-8a63-458a-b7e3-8fb89ff95fc0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-route-controller-manager/route-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:00Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.954009 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.954068 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.954085 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.954109 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:00 crc kubenswrapper[4813]: I1206 15:47:00.954126 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:00Z","lastTransitionTime":"2025-12-06T15:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.057239 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.057332 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.057350 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.057377 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.057394 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:01Z","lastTransitionTime":"2025-12-06T15:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.161141 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.161215 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.161238 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.161300 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.161319 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:01Z","lastTransitionTime":"2025-12-06T15:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.265382 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.265447 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.265463 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.265490 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.265508 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:01Z","lastTransitionTime":"2025-12-06T15:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.369235 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.369340 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.369363 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.369388 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.369406 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:01Z","lastTransitionTime":"2025-12-06T15:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.472591 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.472642 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.472654 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.472671 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.472684 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:01Z","lastTransitionTime":"2025-12-06T15:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.485964 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.486013 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.486035 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.486151 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:01 crc kubenswrapper[4813]: E1206 15:47:01.486158 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:47:01 crc kubenswrapper[4813]: E1206 15:47:01.486313 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:47:01 crc kubenswrapper[4813]: E1206 15:47:01.486469 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:47:01 crc kubenswrapper[4813]: E1206 15:47:01.486582 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.576358 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.576437 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.576470 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.576506 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.576528 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:01Z","lastTransitionTime":"2025-12-06T15:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.680149 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.680212 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.680236 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.680301 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.680329 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:01Z","lastTransitionTime":"2025-12-06T15:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.782891 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.782978 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.782996 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.783022 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.783039 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:01Z","lastTransitionTime":"2025-12-06T15:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.887055 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.887111 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.887128 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.887153 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.887169 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:01Z","lastTransitionTime":"2025-12-06T15:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.990204 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.990240 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.990248 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.990282 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:01 crc kubenswrapper[4813]: I1206 15:47:01.990294 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:01Z","lastTransitionTime":"2025-12-06T15:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.093449 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.093506 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.093524 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.093549 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.093567 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:02Z","lastTransitionTime":"2025-12-06T15:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.196083 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.196137 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.196158 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.196183 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.196200 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:02Z","lastTransitionTime":"2025-12-06T15:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.299339 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.299399 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.299418 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.299443 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.299464 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:02Z","lastTransitionTime":"2025-12-06T15:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.403020 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.403486 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.403679 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.403838 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.403999 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:02Z","lastTransitionTime":"2025-12-06T15:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.506883 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.506966 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.506994 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.507023 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.507042 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:02Z","lastTransitionTime":"2025-12-06T15:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.609823 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.609873 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.609888 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.609910 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.609924 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:02Z","lastTransitionTime":"2025-12-06T15:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.713036 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.713092 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.713107 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.713129 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.713145 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:02Z","lastTransitionTime":"2025-12-06T15:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.815776 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.816095 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.816175 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.816316 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.816473 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:02Z","lastTransitionTime":"2025-12-06T15:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.919895 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.919947 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.919964 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.919987 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:02 crc kubenswrapper[4813]: I1206 15:47:02.920004 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:02Z","lastTransitionTime":"2025-12-06T15:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.023315 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.023360 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.023377 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.023401 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.023420 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:03Z","lastTransitionTime":"2025-12-06T15:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.126457 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.126492 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.126502 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.126518 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.126530 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:03Z","lastTransitionTime":"2025-12-06T15:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.229892 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.229958 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.229977 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.230001 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.230019 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:03Z","lastTransitionTime":"2025-12-06T15:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.333946 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.333986 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.333998 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.334013 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.334025 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:03Z","lastTransitionTime":"2025-12-06T15:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.435986 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.436020 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.436031 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.436045 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.436055 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:03Z","lastTransitionTime":"2025-12-06T15:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.486853 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.486909 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.486872 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.486853 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:03 crc kubenswrapper[4813]: E1206 15:47:03.487017 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:47:03 crc kubenswrapper[4813]: E1206 15:47:03.487106 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:47:03 crc kubenswrapper[4813]: E1206 15:47:03.487228 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:47:03 crc kubenswrapper[4813]: E1206 15:47:03.487342 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.539605 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.539664 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.539674 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.539689 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.539701 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:03Z","lastTransitionTime":"2025-12-06T15:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.643102 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.643150 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.643167 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.643200 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.643217 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:03Z","lastTransitionTime":"2025-12-06T15:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.750426 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.750486 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.750503 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.750526 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.750542 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:03Z","lastTransitionTime":"2025-12-06T15:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.852795 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.852833 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.852844 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.852860 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.852870 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:03Z","lastTransitionTime":"2025-12-06T15:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.955386 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.955420 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.955431 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.955446 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:03 crc kubenswrapper[4813]: I1206 15:47:03.955456 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:03Z","lastTransitionTime":"2025-12-06T15:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.058361 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.058613 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.058737 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.058836 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.058919 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:04Z","lastTransitionTime":"2025-12-06T15:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.161523 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.161563 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.161572 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.161590 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.161608 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:04Z","lastTransitionTime":"2025-12-06T15:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.264131 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.264174 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.264184 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.264205 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.264216 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:04Z","lastTransitionTime":"2025-12-06T15:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.366748 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.366777 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.366788 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.366801 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.366809 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:04Z","lastTransitionTime":"2025-12-06T15:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.469376 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.469410 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.469421 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.469437 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.469448 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:04Z","lastTransitionTime":"2025-12-06T15:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.575569 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.575608 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.575617 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.575632 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.575641 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:04Z","lastTransitionTime":"2025-12-06T15:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.678996 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.679042 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.679053 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.679068 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.679078 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:04Z","lastTransitionTime":"2025-12-06T15:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.781112 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.781375 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.781528 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.781666 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.781758 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:04Z","lastTransitionTime":"2025-12-06T15:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.885413 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.885623 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.885700 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.885765 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.885825 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:04Z","lastTransitionTime":"2025-12-06T15:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.988038 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.988381 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.988614 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.988770 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:04 crc kubenswrapper[4813]: I1206 15:47:04.988909 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:04Z","lastTransitionTime":"2025-12-06T15:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.091597 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.091637 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.091649 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.091664 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.091675 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:05Z","lastTransitionTime":"2025-12-06T15:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.194497 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.194537 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.194548 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.194566 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.194578 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:05Z","lastTransitionTime":"2025-12-06T15:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.297323 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.297596 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.297610 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.297627 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.297639 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:05Z","lastTransitionTime":"2025-12-06T15:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.400512 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.400586 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.400609 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.400640 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.400662 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:05Z","lastTransitionTime":"2025-12-06T15:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.486175 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:05 crc kubenswrapper[4813]: E1206 15:47:05.486323 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.486354 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.486379 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:05 crc kubenswrapper[4813]: E1206 15:47:05.486469 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:47:05 crc kubenswrapper[4813]: E1206 15:47:05.486572 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.487111 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:05 crc kubenswrapper[4813]: E1206 15:47:05.487446 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.487478 4813 scope.go:117] "RemoveContainer" containerID="0d6b3583626fb73a7ee10d9ea994c5dc388215022154a5a51777e14a8195cc0c" Dec 06 15:47:05 crc kubenswrapper[4813]: E1206 15:47:05.490696 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xdt4d_openshift-ovn-kubernetes(0e54e31f-e9f8-4dd5-8738-c785df002c93)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.505798 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.505872 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.505895 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.505925 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.505948 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:05Z","lastTransitionTime":"2025-12-06T15:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.508112 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:05Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.522111 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:05Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.536455 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:05Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.554300 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:05Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.566823 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:05Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.591746 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d6b3583626fb73a7ee10d9ea994c5dc388215022154a5a51777e14a8195cc0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d6b3583626fb73a7ee10d9ea994c5dc388215022154a5a51777e14a8195cc0c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T15:46:49Z\\\",\\\"message\\\":\\\"ontroller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:49Z is after 2025-08-24T17:21:41Z]\\\\nI1206 15:46:49.421493 6320 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"18746a4d-8a63-458a-b7e3-8fb89ff95fc0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-route-controller-manager/route-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xdt4d_openshift-ovn-kubernetes(0e54e31f-e9f8-4dd5-8738-c785df002c93)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:05Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.605693 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:05Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.609088 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.609248 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.609408 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.609546 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.609662 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:05Z","lastTransitionTime":"2025-12-06T15:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.622540 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"012e3f9a-67f1-45f2-8e57-76928bb4d5c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a77b5cb2e8c50b39c3369f156a7800603e7d8bf3d6a544ad0f3a6814361338f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2f1dd12351d9f1ce3ccc9711361cf346d9aa1c4cb5d971020e242c80e557eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8326b53a1a99f7cbd09d55008e9c1b4087993a25ae4b50fafc70e4c7b99138d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9471d3db940ee579d5b60cf1f7750c2c265a8f8e549aaa7ed60e82982cf80d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9471d3db940ee579d5b60cf1f7750c2c265a8f8e549aaa7ed60e82982cf80d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:05Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.635922 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:05Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.649744 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:05Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.670223 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:05Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.683104 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:05Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.697192 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a379c66da38cd6d16eb0fdc5941e3fd5d4127b7cce84fe2bde55918ef2c99c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:05Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.712394 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f228acd2a1febab7d6991054e11cf509ff83ab26e537db82ab49688144a4a233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:05Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.712683 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.712702 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.712712 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.712741 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.712750 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:05Z","lastTransitionTime":"2025-12-06T15:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.726297 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:05Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.743549 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:05Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.753665 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ab1d9f15bba233c4bbd04ece144a5a8119c168abaeee7d4544269b55ad58ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b14a8cf06c49b6922327f25905466f86e1a0fd0a0b852f751bac6f8cd7db9e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bd4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:05Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.762528 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kzbhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18a94103-5ed7-4e27-a58d-7c989b0f70f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kzbhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:05Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.815151 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.815318 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.815400 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.815490 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.815570 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:05Z","lastTransitionTime":"2025-12-06T15:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.917879 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.917930 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.917939 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.917956 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:05 crc kubenswrapper[4813]: I1206 15:47:05.917967 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:05Z","lastTransitionTime":"2025-12-06T15:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.020581 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.020936 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.021191 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.021486 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.021703 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:06Z","lastTransitionTime":"2025-12-06T15:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.124447 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.124479 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.124490 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.124506 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.124515 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:06Z","lastTransitionTime":"2025-12-06T15:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.226980 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.227220 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.227298 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.227368 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.227436 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:06Z","lastTransitionTime":"2025-12-06T15:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.330304 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.330342 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.330351 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.330368 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.330380 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:06Z","lastTransitionTime":"2025-12-06T15:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.415214 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.415296 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.415314 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.415337 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.415355 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:06Z","lastTransitionTime":"2025-12-06T15:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:06 crc kubenswrapper[4813]: E1206 15:47:06.439394 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:06Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.445550 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.445851 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.446019 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.446143 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.446289 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:06Z","lastTransitionTime":"2025-12-06T15:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:06 crc kubenswrapper[4813]: E1206 15:47:06.465902 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:06Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.471215 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.471326 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.471348 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.471404 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.471423 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:06Z","lastTransitionTime":"2025-12-06T15:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:06 crc kubenswrapper[4813]: E1206 15:47:06.492097 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:06Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.499428 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.499474 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.499491 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.499517 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.499536 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:06Z","lastTransitionTime":"2025-12-06T15:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:06 crc kubenswrapper[4813]: E1206 15:47:06.513922 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:06Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.518021 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.518052 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.518062 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.518077 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.518088 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:06Z","lastTransitionTime":"2025-12-06T15:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:06 crc kubenswrapper[4813]: E1206 15:47:06.530911 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:06Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:06 crc kubenswrapper[4813]: E1206 15:47:06.531058 4813 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.532593 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.532700 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.532716 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.532730 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.532740 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:06Z","lastTransitionTime":"2025-12-06T15:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.635481 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.635745 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.635829 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.635918 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.635999 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:06Z","lastTransitionTime":"2025-12-06T15:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.738730 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.738765 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.738774 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.738791 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.738802 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:06Z","lastTransitionTime":"2025-12-06T15:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.841100 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.841134 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.841143 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.841157 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.841168 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:06Z","lastTransitionTime":"2025-12-06T15:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.944866 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.944897 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.944906 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.944941 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:06 crc kubenswrapper[4813]: I1206 15:47:06.944950 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:06Z","lastTransitionTime":"2025-12-06T15:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.048413 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.048473 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.048492 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.048521 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.048543 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:07Z","lastTransitionTime":"2025-12-06T15:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.151854 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.151915 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.151929 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.151946 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.151958 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:07Z","lastTransitionTime":"2025-12-06T15:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.255291 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.255405 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.255482 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.255561 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.255623 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:07Z","lastTransitionTime":"2025-12-06T15:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.359092 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.359165 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.359184 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.359210 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.359228 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:07Z","lastTransitionTime":"2025-12-06T15:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.402106 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18a94103-5ed7-4e27-a58d-7c989b0f70f3-metrics-certs\") pod \"network-metrics-daemon-kzbhb\" (UID: \"18a94103-5ed7-4e27-a58d-7c989b0f70f3\") " pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:07 crc kubenswrapper[4813]: E1206 15:47:07.402373 4813 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 15:47:07 crc kubenswrapper[4813]: E1206 15:47:07.402503 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18a94103-5ed7-4e27-a58d-7c989b0f70f3-metrics-certs podName:18a94103-5ed7-4e27-a58d-7c989b0f70f3 nodeName:}" failed. No retries permitted until 2025-12-06 15:47:39.402476475 +0000 UTC m=+99.293356061 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/18a94103-5ed7-4e27-a58d-7c989b0f70f3-metrics-certs") pod "network-metrics-daemon-kzbhb" (UID: "18a94103-5ed7-4e27-a58d-7c989b0f70f3") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.462139 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.462644 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.462749 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.462851 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.462923 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:07Z","lastTransitionTime":"2025-12-06T15:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.486216 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.486371 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.486383 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.486241 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:07 crc kubenswrapper[4813]: E1206 15:47:07.486611 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:47:07 crc kubenswrapper[4813]: E1206 15:47:07.486742 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:47:07 crc kubenswrapper[4813]: E1206 15:47:07.486850 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:47:07 crc kubenswrapper[4813]: E1206 15:47:07.486967 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.565698 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.565875 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.566046 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.566209 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.566376 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:07Z","lastTransitionTime":"2025-12-06T15:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.668835 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.668899 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.668917 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.668944 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.668962 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:07Z","lastTransitionTime":"2025-12-06T15:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.771955 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.772000 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.772011 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.772044 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.772056 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:07Z","lastTransitionTime":"2025-12-06T15:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.875150 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.875197 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.875212 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.875231 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.875242 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:07Z","lastTransitionTime":"2025-12-06T15:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.979003 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.979070 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.979086 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.979116 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:07 crc kubenswrapper[4813]: I1206 15:47:07.979137 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:07Z","lastTransitionTime":"2025-12-06T15:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.082450 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.082504 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.082516 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.082534 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.082547 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:08Z","lastTransitionTime":"2025-12-06T15:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.185682 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.185726 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.185739 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.185755 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.185767 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:08Z","lastTransitionTime":"2025-12-06T15:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.289153 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.289210 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.289227 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.289254 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.289299 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:08Z","lastTransitionTime":"2025-12-06T15:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.392721 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.392771 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.392784 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.392804 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.392822 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:08Z","lastTransitionTime":"2025-12-06T15:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.495136 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.495197 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.495215 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.495239 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.495256 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:08Z","lastTransitionTime":"2025-12-06T15:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.597754 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.597845 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.597860 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.597880 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.597894 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:08Z","lastTransitionTime":"2025-12-06T15:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.701092 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.701138 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.701149 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.701168 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.701179 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:08Z","lastTransitionTime":"2025-12-06T15:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.804636 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.804686 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.804696 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.804713 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.804726 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:08Z","lastTransitionTime":"2025-12-06T15:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.907287 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.907352 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.907368 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.907387 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:08 crc kubenswrapper[4813]: I1206 15:47:08.907399 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:08Z","lastTransitionTime":"2025-12-06T15:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.010495 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.010567 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.010581 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.010602 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.010633 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:09Z","lastTransitionTime":"2025-12-06T15:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.114253 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.114329 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.114347 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.114372 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.114393 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:09Z","lastTransitionTime":"2025-12-06T15:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.216702 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.216775 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.216790 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.216809 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.216824 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:09Z","lastTransitionTime":"2025-12-06T15:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.319789 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.319835 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.319846 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.319867 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.319879 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:09Z","lastTransitionTime":"2025-12-06T15:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.423370 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.423423 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.423434 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.423452 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.423464 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:09Z","lastTransitionTime":"2025-12-06T15:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.486687 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.486687 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.486711 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.486848 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:09 crc kubenswrapper[4813]: E1206 15:47:09.486920 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:47:09 crc kubenswrapper[4813]: E1206 15:47:09.487064 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:47:09 crc kubenswrapper[4813]: E1206 15:47:09.487097 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:47:09 crc kubenswrapper[4813]: E1206 15:47:09.487167 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.525653 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.525692 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.525705 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.525723 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.525734 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:09Z","lastTransitionTime":"2025-12-06T15:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.627869 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.627913 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.627926 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.627938 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.627947 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:09Z","lastTransitionTime":"2025-12-06T15:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.730756 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.730800 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.730815 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.730835 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.730848 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:09Z","lastTransitionTime":"2025-12-06T15:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.833432 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.833484 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.833498 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.833521 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.833539 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:09Z","lastTransitionTime":"2025-12-06T15:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.936317 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.936360 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.936371 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.936427 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:09 crc kubenswrapper[4813]: I1206 15:47:09.936440 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:09Z","lastTransitionTime":"2025-12-06T15:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.039945 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.039998 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.040013 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.040031 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.040044 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:10Z","lastTransitionTime":"2025-12-06T15:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.142957 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.143000 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.143013 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.143034 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.143047 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:10Z","lastTransitionTime":"2025-12-06T15:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.245446 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.245525 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.245551 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.245582 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.245606 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:10Z","lastTransitionTime":"2025-12-06T15:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.349289 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.349334 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.349350 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.349367 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.349383 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:10Z","lastTransitionTime":"2025-12-06T15:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.452303 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.452399 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.452423 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.452453 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.452473 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:10Z","lastTransitionTime":"2025-12-06T15:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.509182 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:10Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.529007 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:10Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.531114 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gl495_856bb45b-525c-4ef0-bf7c-0691cf54b342/kube-multus/0.log" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.531222 4813 generic.go:334] "Generic (PLEG): container finished" podID="856bb45b-525c-4ef0-bf7c-0691cf54b342" containerID="e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c" exitCode=1 Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.531315 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gl495" event={"ID":"856bb45b-525c-4ef0-bf7c-0691cf54b342","Type":"ContainerDied","Data":"e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c"} Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.532013 4813 scope.go:117] "RemoveContainer" containerID="e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.548459 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a379c66da38cd6d16eb0fdc5941e3fd5d4127b7cce84fe2bde55918ef2c99c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:10Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.564434 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.564652 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.564713 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.564771 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.564790 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:10Z","lastTransitionTime":"2025-12-06T15:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.581967 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f228acd2a1febab7d6991054e11cf509ff83ab26e537db82ab49688144a4a233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:10Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.604647 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:10Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.626447 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:10Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.637941 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ab1d9f15bba233c4bbd04ece144a5a8119c168abaeee7d4544269b55ad58ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b14a8cf06c49b6922327f25905466f86e1a0fd0a0b852f751bac6f8cd7db9e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bd4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:10Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.651662 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kzbhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18a94103-5ed7-4e27-a58d-7c989b0f70f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kzbhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:10Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.667418 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.667459 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.667471 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.667488 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.667500 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:10Z","lastTransitionTime":"2025-12-06T15:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.667712 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d6b3583626fb73a7ee10d9ea994c5dc388215022154a5a51777e14a8195cc0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d6b3583626fb73a7ee10d9ea994c5dc388215022154a5a51777e14a8195cc0c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T15:46:49Z\\\",\\\"message\\\":\\\"ontroller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:49Z is after 2025-08-24T17:21:41Z]\\\\nI1206 15:46:49.421493 6320 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"18746a4d-8a63-458a-b7e3-8fb89ff95fc0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-route-controller-manager/route-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xdt4d_openshift-ovn-kubernetes(0e54e31f-e9f8-4dd5-8738-c785df002c93)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:10Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.680373 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:10Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.698516 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:10Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.741848 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:10Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.768562 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:10Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.770873 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.770936 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.770948 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.770986 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.771002 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:10Z","lastTransitionTime":"2025-12-06T15:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.781564 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:10Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.791129 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:10Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.801615 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"012e3f9a-67f1-45f2-8e57-76928bb4d5c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a77b5cb2e8c50b39c3369f156a7800603e7d8bf3d6a544ad0f3a6814361338f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2f1dd12351d9f1ce3ccc9711361cf346d9aa1c4cb5d971020e242c80e557eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8326b53a1a99f7cbd09d55008e9c1b4087993a25ae4b50fafc70e4c7b99138d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9471d3db940ee579d5b60cf1f7750c2c265a8f8e549aaa7ed60e82982cf80d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9471d3db940ee579d5b60cf1f7750c2c265a8f8e549aaa7ed60e82982cf80d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:10Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.814008 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:10Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.825140 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:10Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.839291 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:10Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.853792 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:10Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.865421 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:10Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.873684 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.873739 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.873753 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.873776 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.873788 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:10Z","lastTransitionTime":"2025-12-06T15:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.880512 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:10Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.889586 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:10Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.906861 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d6b3583626fb73a7ee10d9ea994c5dc388215022154a5a51777e14a8195cc0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d6b3583626fb73a7ee10d9ea994c5dc388215022154a5a51777e14a8195cc0c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T15:46:49Z\\\",\\\"message\\\":\\\"ontroller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:49Z is after 2025-08-24T17:21:41Z]\\\\nI1206 15:46:49.421493 6320 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"18746a4d-8a63-458a-b7e3-8fb89ff95fc0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-route-controller-manager/route-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xdt4d_openshift-ovn-kubernetes(0e54e31f-e9f8-4dd5-8738-c785df002c93)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:10Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.919422 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:10Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.936757 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"012e3f9a-67f1-45f2-8e57-76928bb4d5c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a77b5cb2e8c50b39c3369f156a7800603e7d8bf3d6a544ad0f3a6814361338f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2f1dd12351d9f1ce3ccc9711361cf346d9aa1c4cb5d971020e242c80e557eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8326b53a1a99f7cbd09d55008e9c1b4087993a25ae4b50fafc70e4c7b99138d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9471d3db940ee579d5b60cf1f7750c2c265a8f8e549aaa7ed60e82982cf80d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9471d3db940ee579d5b60cf1f7750c2c265a8f8e549aaa7ed60e82982cf80d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:10Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.948908 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:10Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.962945 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:10Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.976971 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.977237 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.977335 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.977416 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.977488 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:10Z","lastTransitionTime":"2025-12-06T15:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.979233 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:10Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:10 crc kubenswrapper[4813]: I1206 15:47:10.994725 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:10Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.007673 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a379c66da38cd6d16eb0fdc5941e3fd5d4127b7cce84fe2bde55918ef2c99c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:11Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.023308 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f228acd2a1febab7d6991054e11cf509ff83ab26e537db82ab49688144a4a233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:11Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.044433 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T15:47:09Z\\\",\\\"message\\\":\\\"2025-12-06T15:46:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5f77eacc-42b0-4fed-a916-5e52a3f47837\\\\n2025-12-06T15:46:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5f77eacc-42b0-4fed-a916-5e52a3f47837 to /host/opt/cni/bin/\\\\n2025-12-06T15:46:24Z [verbose] multus-daemon started\\\\n2025-12-06T15:46:24Z [verbose] Readiness Indicator file check\\\\n2025-12-06T15:47:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:11Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.075515 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:11Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.085854 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.086076 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.086323 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.086402 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.086460 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:11Z","lastTransitionTime":"2025-12-06T15:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.090410 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ab1d9f15bba233c4bbd04ece144a5a8119c168abaeee7d4544269b55ad58ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b14a8cf06c49b6922327f25905466f86e1a0fd0a0b852f751bac6f8cd7db9e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bd4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:11Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.105975 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kzbhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18a94103-5ed7-4e27-a58d-7c989b0f70f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kzbhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:11Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.189609 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.189921 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.190012 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.190150 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.190229 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:11Z","lastTransitionTime":"2025-12-06T15:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.293478 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.293849 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.294451 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.295304 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.295358 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:11Z","lastTransitionTime":"2025-12-06T15:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.398780 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.399080 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.399143 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.399220 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.399347 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:11Z","lastTransitionTime":"2025-12-06T15:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.485917 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.486231 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.486142 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.486014 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:11 crc kubenswrapper[4813]: E1206 15:47:11.486608 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:47:11 crc kubenswrapper[4813]: E1206 15:47:11.486858 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:47:11 crc kubenswrapper[4813]: E1206 15:47:11.486988 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:47:11 crc kubenswrapper[4813]: E1206 15:47:11.487187 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.503444 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.503568 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.503676 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.503768 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.503858 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:11Z","lastTransitionTime":"2025-12-06T15:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.538311 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gl495_856bb45b-525c-4ef0-bf7c-0691cf54b342/kube-multus/0.log" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.539300 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gl495" event={"ID":"856bb45b-525c-4ef0-bf7c-0691cf54b342","Type":"ContainerStarted","Data":"7001bdf05c6acf52c9be35bd63ae103390bf04d2a00a41bf08d376dcdc497338"} Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.557599 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a379c66da38cd6d16eb0fdc5941e3fd5d4127b7cce84fe2bde55918ef2c99c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:11Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.580133 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f228acd2a1febab7d6991054e11cf509ff83ab26e537db82ab49688144a4a233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:11Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.599883 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7001bdf05c6acf52c9be35bd63ae103390bf04d2a00a41bf08d376dcdc497338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T15:47:09Z\\\",\\\"message\\\":\\\"2025-12-06T15:46:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5f77eacc-42b0-4fed-a916-5e52a3f47837\\\\n2025-12-06T15:46:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5f77eacc-42b0-4fed-a916-5e52a3f47837 to /host/opt/cni/bin/\\\\n2025-12-06T15:46:24Z [verbose] multus-daemon started\\\\n2025-12-06T15:46:24Z [verbose] Readiness Indicator file check\\\\n2025-12-06T15:47:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:47:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:11Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.608959 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.609294 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.609323 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.609349 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.609366 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:11Z","lastTransitionTime":"2025-12-06T15:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.613502 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:11Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.626415 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:11Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.638294 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kzbhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18a94103-5ed7-4e27-a58d-7c989b0f70f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kzbhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:11Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.656669 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:11Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.668021 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ab1d9f15bba233c4bbd04ece144a5a8119c168abaeee7d4544269b55ad58ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b14a8cf06c49b6922327f25905466f86e1a0fd0a0b852f751bac6f8cd7db9e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bd4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:11Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.681296 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:11Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.693668 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:11Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.708174 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:11Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.712307 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.712366 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.712380 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.712400 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.712412 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:11Z","lastTransitionTime":"2025-12-06T15:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.736256 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d6b3583626fb73a7ee10d9ea994c5dc388215022154a5a51777e14a8195cc0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d6b3583626fb73a7ee10d9ea994c5dc388215022154a5a51777e14a8195cc0c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T15:46:49Z\\\",\\\"message\\\":\\\"ontroller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:49Z is after 2025-08-24T17:21:41Z]\\\\nI1206 15:46:49.421493 6320 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"18746a4d-8a63-458a-b7e3-8fb89ff95fc0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-route-controller-manager/route-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xdt4d_openshift-ovn-kubernetes(0e54e31f-e9f8-4dd5-8738-c785df002c93)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:11Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.810519 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:11Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.815802 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.815836 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.815847 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.815863 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.815873 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:11Z","lastTransitionTime":"2025-12-06T15:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.821756 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:11Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.833069 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:11Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.843176 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:11Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.852529 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:11Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.861786 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"012e3f9a-67f1-45f2-8e57-76928bb4d5c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a77b5cb2e8c50b39c3369f156a7800603e7d8bf3d6a544ad0f3a6814361338f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2f1dd12351d9f1ce3ccc9711361cf346d9aa1c4cb5d971020e242c80e557eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8326b53a1a99f7cbd09d55008e9c1b4087993a25ae4b50fafc70e4c7b99138d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9471d3db940ee579d5b60cf1f7750c2c265a8f8e549aaa7ed60e82982cf80d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9471d3db940ee579d5b60cf1f7750c2c265a8f8e549aaa7ed60e82982cf80d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:11Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.917704 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.917734 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.917744 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.917756 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:11 crc kubenswrapper[4813]: I1206 15:47:11.917765 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:11Z","lastTransitionTime":"2025-12-06T15:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.019843 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.019885 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.019898 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.019915 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.019934 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:12Z","lastTransitionTime":"2025-12-06T15:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.122378 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.122424 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.122443 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.122465 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.122481 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:12Z","lastTransitionTime":"2025-12-06T15:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.225492 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.225533 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.225543 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.225557 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.225567 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:12Z","lastTransitionTime":"2025-12-06T15:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.328069 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.328113 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.328127 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.328146 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.328159 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:12Z","lastTransitionTime":"2025-12-06T15:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.431312 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.431364 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.431377 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.431395 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.431406 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:12Z","lastTransitionTime":"2025-12-06T15:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.494537 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.533544 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.533581 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.533591 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.533605 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.533615 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:12Z","lastTransitionTime":"2025-12-06T15:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.635448 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.635501 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.635524 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.635554 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.635576 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:12Z","lastTransitionTime":"2025-12-06T15:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.739959 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.739998 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.740006 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.740021 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.740032 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:12Z","lastTransitionTime":"2025-12-06T15:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.841979 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.842013 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.842025 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.842040 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.842051 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:12Z","lastTransitionTime":"2025-12-06T15:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.945462 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.945504 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.945514 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.945530 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:12 crc kubenswrapper[4813]: I1206 15:47:12.945540 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:12Z","lastTransitionTime":"2025-12-06T15:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.048697 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.048730 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.048740 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.048756 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.048767 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:13Z","lastTransitionTime":"2025-12-06T15:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.151567 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.151620 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.151633 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.151649 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.151661 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:13Z","lastTransitionTime":"2025-12-06T15:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.254559 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.254606 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.254624 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.254649 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.254668 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:13Z","lastTransitionTime":"2025-12-06T15:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.358187 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.358254 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.358287 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.358308 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.358326 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:13Z","lastTransitionTime":"2025-12-06T15:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.462024 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.462308 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.462513 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.462658 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.462781 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:13Z","lastTransitionTime":"2025-12-06T15:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.486617 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.486658 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.486613 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:13 crc kubenswrapper[4813]: E1206 15:47:13.486731 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:47:13 crc kubenswrapper[4813]: E1206 15:47:13.486947 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:47:13 crc kubenswrapper[4813]: E1206 15:47:13.487173 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.487567 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:13 crc kubenswrapper[4813]: E1206 15:47:13.487862 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.565800 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.565857 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.565873 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.565898 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.565917 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:13Z","lastTransitionTime":"2025-12-06T15:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.669150 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.669280 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.669301 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.669319 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.669331 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:13Z","lastTransitionTime":"2025-12-06T15:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.771962 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.771998 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.772009 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.772025 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.772038 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:13Z","lastTransitionTime":"2025-12-06T15:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.874622 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.874708 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.874739 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.874772 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.874794 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:13Z","lastTransitionTime":"2025-12-06T15:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.976974 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.977007 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.977017 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.977032 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:13 crc kubenswrapper[4813]: I1206 15:47:13.977043 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:13Z","lastTransitionTime":"2025-12-06T15:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.079363 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.079411 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.079427 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.079449 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.079466 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:14Z","lastTransitionTime":"2025-12-06T15:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.182606 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.182645 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.182663 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.182683 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.182700 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:14Z","lastTransitionTime":"2025-12-06T15:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.286744 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.286806 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.286826 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.286852 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.286872 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:14Z","lastTransitionTime":"2025-12-06T15:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.390506 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.390555 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.390568 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.390588 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.390601 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:14Z","lastTransitionTime":"2025-12-06T15:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.493683 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.493758 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.493784 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.493821 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.493846 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:14Z","lastTransitionTime":"2025-12-06T15:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.597698 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.597766 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.597790 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.597815 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.597836 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:14Z","lastTransitionTime":"2025-12-06T15:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.701806 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.702078 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.702103 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.702138 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.702162 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:14Z","lastTransitionTime":"2025-12-06T15:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.805801 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.805867 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.805880 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.805903 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.805917 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:14Z","lastTransitionTime":"2025-12-06T15:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.908866 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.908909 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.908921 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.908940 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:14 crc kubenswrapper[4813]: I1206 15:47:14.908952 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:14Z","lastTransitionTime":"2025-12-06T15:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.012824 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.012868 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.012881 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.012902 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.012918 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:15Z","lastTransitionTime":"2025-12-06T15:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.116603 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.116707 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.116722 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.116749 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.116764 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:15Z","lastTransitionTime":"2025-12-06T15:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.220955 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.221019 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.221032 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.221056 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.221069 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:15Z","lastTransitionTime":"2025-12-06T15:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.324756 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.324806 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.324823 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.324847 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.324864 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:15Z","lastTransitionTime":"2025-12-06T15:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.428886 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.429069 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.429129 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.429161 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.429212 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:15Z","lastTransitionTime":"2025-12-06T15:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.486747 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.486980 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:15 crc kubenswrapper[4813]: E1206 15:47:15.487138 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.487345 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.487352 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:15 crc kubenswrapper[4813]: E1206 15:47:15.487666 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:47:15 crc kubenswrapper[4813]: E1206 15:47:15.487512 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:47:15 crc kubenswrapper[4813]: E1206 15:47:15.488002 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.533574 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.533642 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.533668 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.533703 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.533727 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:15Z","lastTransitionTime":"2025-12-06T15:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.637727 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.637802 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.637825 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.637858 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.637882 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:15Z","lastTransitionTime":"2025-12-06T15:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.741414 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.741487 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.741509 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.741541 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.741562 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:15Z","lastTransitionTime":"2025-12-06T15:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.847314 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.847471 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.847487 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.847550 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.847566 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:15Z","lastTransitionTime":"2025-12-06T15:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.950249 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.950308 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.950319 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.950336 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:15 crc kubenswrapper[4813]: I1206 15:47:15.950347 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:15Z","lastTransitionTime":"2025-12-06T15:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.054840 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.054911 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.054924 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.054943 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.054956 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:16Z","lastTransitionTime":"2025-12-06T15:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.158229 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.158336 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.158348 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.158390 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.158403 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:16Z","lastTransitionTime":"2025-12-06T15:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.261733 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.261805 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.261823 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.261854 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.261872 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:16Z","lastTransitionTime":"2025-12-06T15:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.364838 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.364909 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.364923 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.364948 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.364963 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:16Z","lastTransitionTime":"2025-12-06T15:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.467344 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.467423 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.467438 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.467459 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.467473 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:16Z","lastTransitionTime":"2025-12-06T15:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.487438 4813 scope.go:117] "RemoveContainer" containerID="0d6b3583626fb73a7ee10d9ea994c5dc388215022154a5a51777e14a8195cc0c" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.570518 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.570970 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.570990 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.571014 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.571033 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:16Z","lastTransitionTime":"2025-12-06T15:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.673677 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.673743 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.673767 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.673799 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.673824 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:16Z","lastTransitionTime":"2025-12-06T15:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.777150 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.777229 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.777341 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.777385 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.777461 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:16Z","lastTransitionTime":"2025-12-06T15:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.838371 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.838409 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.838449 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.838518 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.838532 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:16Z","lastTransitionTime":"2025-12-06T15:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:16 crc kubenswrapper[4813]: E1206 15:47:16.859972 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:16Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.865162 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.865516 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.865697 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.865865 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.866010 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:16Z","lastTransitionTime":"2025-12-06T15:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:16 crc kubenswrapper[4813]: E1206 15:47:16.884917 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:16Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.889609 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.889661 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.889679 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.889705 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.889723 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:16Z","lastTransitionTime":"2025-12-06T15:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:16 crc kubenswrapper[4813]: E1206 15:47:16.904859 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:16Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.911246 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.911330 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.911354 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.911387 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.911407 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:16Z","lastTransitionTime":"2025-12-06T15:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:16 crc kubenswrapper[4813]: E1206 15:47:16.924197 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:16Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.929097 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.929189 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.929215 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.929254 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.929344 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:16Z","lastTransitionTime":"2025-12-06T15:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:16 crc kubenswrapper[4813]: E1206 15:47:16.946943 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:16Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:16 crc kubenswrapper[4813]: E1206 15:47:16.947307 4813 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.949722 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.949801 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.949826 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.949853 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:16 crc kubenswrapper[4813]: I1206 15:47:16.949876 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:16Z","lastTransitionTime":"2025-12-06T15:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.057852 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.057901 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.057913 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.057932 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.057943 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:17Z","lastTransitionTime":"2025-12-06T15:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.160955 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.161001 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.161011 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.161029 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.161038 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:17Z","lastTransitionTime":"2025-12-06T15:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.272436 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.272470 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.272480 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.272499 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.272511 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:17Z","lastTransitionTime":"2025-12-06T15:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.375195 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.375239 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.375253 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.375289 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.375301 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:17Z","lastTransitionTime":"2025-12-06T15:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.489039 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:17 crc kubenswrapper[4813]: E1206 15:47:17.489189 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.489274 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.489340 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:17 crc kubenswrapper[4813]: E1206 15:47:17.489431 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:47:17 crc kubenswrapper[4813]: E1206 15:47:17.489541 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.490342 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:17 crc kubenswrapper[4813]: E1206 15:47:17.490873 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.491778 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.491803 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.491813 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.491827 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.491838 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:17Z","lastTransitionTime":"2025-12-06T15:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.563459 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xdt4d_0e54e31f-e9f8-4dd5-8738-c785df002c93/ovnkube-controller/2.log" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.566660 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" event={"ID":"0e54e31f-e9f8-4dd5-8738-c785df002c93","Type":"ContainerStarted","Data":"6b654f2820e7e70a18617e577d1dffa10f2e6e786d8e8a2483493d1e2cbf32e6"} Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.568387 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.586666 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:17Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.605576 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:17Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.621153 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:17Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.645894 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b654f2820e7e70a18617e577d1dffa10f2e6e786d8e8a2483493d1e2cbf32e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d6b3583626fb73a7ee10d9ea994c5dc388215022154a5a51777e14a8195cc0c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T15:46:49Z\\\",\\\"message\\\":\\\"ontroller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:49Z is after 2025-08-24T17:21:41Z]\\\\nI1206 15:46:49.421493 6320 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"18746a4d-8a63-458a-b7e3-8fb89ff95fc0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-route-controller-manager/route-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:17Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.652749 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.652884 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.652962 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.653045 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.653108 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:17Z","lastTransitionTime":"2025-12-06T15:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.664850 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:17Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.684783 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:17Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.707450 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:17Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.734108 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:17Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.753763 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:17Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.755908 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.756126 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.756324 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.756482 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.756628 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:17Z","lastTransitionTime":"2025-12-06T15:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.772754 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"393f3b4b-ef95-4f70-af75-9fa123e0d52b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3d7272d34a92fa558ed95d304ebaf3f6f0e79e5d60538187542d201250d93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e297ada3f58d22f81f4014c0d47f04b0fe2e68c1e032e9b17e85907f27f9a6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e297ada3f58d22f81f4014c0d47f04b0fe2e68c1e032e9b17e85907f27f9a6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:17Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.791694 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"012e3f9a-67f1-45f2-8e57-76928bb4d5c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a77b5cb2e8c50b39c3369f156a7800603e7d8bf3d6a544ad0f3a6814361338f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2f1dd12351d9f1ce3ccc9711361cf346d9aa1c4cb5d971020e242c80e557eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8326b53a1a99f7cbd09d55008e9c1b4087993a25ae4b50fafc70e4c7b99138d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9471d3db940ee579d5b60cf1f7750c2c265a8f8e549aaa7ed60e82982cf80d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9471d3db940ee579d5b60cf1f7750c2c265a8f8e549aaa7ed60e82982cf80d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:17Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.814165 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a379c66da38cd6d16eb0fdc5941e3fd5d4127b7cce84fe2bde55918ef2c99c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:17Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.834155 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f228acd2a1febab7d6991054e11cf509ff83ab26e537db82ab49688144a4a233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:17Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.859937 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.859984 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.860002 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.860022 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.860040 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:17Z","lastTransitionTime":"2025-12-06T15:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.915660 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7001bdf05c6acf52c9be35bd63ae103390bf04d2a00a41bf08d376dcdc497338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T15:47:09Z\\\",\\\"message\\\":\\\"2025-12-06T15:46:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5f77eacc-42b0-4fed-a916-5e52a3f47837\\\\n2025-12-06T15:46:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5f77eacc-42b0-4fed-a916-5e52a3f47837 to /host/opt/cni/bin/\\\\n2025-12-06T15:46:24Z [verbose] multus-daemon started\\\\n2025-12-06T15:46:24Z [verbose] Readiness Indicator file check\\\\n2025-12-06T15:47:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:47:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:17Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.963790 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.963831 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.963844 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.963863 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.963881 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:17Z","lastTransitionTime":"2025-12-06T15:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.976500 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:17Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:17 crc kubenswrapper[4813]: I1206 15:47:17.990442 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:17Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.004108 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kzbhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18a94103-5ed7-4e27-a58d-7c989b0f70f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kzbhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:18Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.034521 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:18Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.052851 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ab1d9f15bba233c4bbd04ece144a5a8119c168abaeee7d4544269b55ad58ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b14a8cf06c49b6922327f25905466f86e1a0fd0a0b852f751bac6f8cd7db9e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bd4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:18Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.067192 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.067243 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.067257 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.067291 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.067304 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:18Z","lastTransitionTime":"2025-12-06T15:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.169437 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.169908 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.170022 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.170138 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.170237 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:18Z","lastTransitionTime":"2025-12-06T15:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.273751 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.273821 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.273840 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.273872 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.273889 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:18Z","lastTransitionTime":"2025-12-06T15:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.377194 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.377300 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.377323 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.377354 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.377374 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:18Z","lastTransitionTime":"2025-12-06T15:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.481647 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.481724 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.481742 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.481768 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.481786 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:18Z","lastTransitionTime":"2025-12-06T15:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.574339 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xdt4d_0e54e31f-e9f8-4dd5-8738-c785df002c93/ovnkube-controller/3.log" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.575417 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xdt4d_0e54e31f-e9f8-4dd5-8738-c785df002c93/ovnkube-controller/2.log" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.579566 4813 generic.go:334] "Generic (PLEG): container finished" podID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerID="6b654f2820e7e70a18617e577d1dffa10f2e6e786d8e8a2483493d1e2cbf32e6" exitCode=1 Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.579641 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" event={"ID":"0e54e31f-e9f8-4dd5-8738-c785df002c93","Type":"ContainerDied","Data":"6b654f2820e7e70a18617e577d1dffa10f2e6e786d8e8a2483493d1e2cbf32e6"} Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.579716 4813 scope.go:117] "RemoveContainer" containerID="0d6b3583626fb73a7ee10d9ea994c5dc388215022154a5a51777e14a8195cc0c" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.580873 4813 scope.go:117] "RemoveContainer" containerID="6b654f2820e7e70a18617e577d1dffa10f2e6e786d8e8a2483493d1e2cbf32e6" Dec 06 15:47:18 crc kubenswrapper[4813]: E1206 15:47:18.581140 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xdt4d_openshift-ovn-kubernetes(0e54e31f-e9f8-4dd5-8738-c785df002c93)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.595032 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.595075 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.595091 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.595114 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.595130 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:18Z","lastTransitionTime":"2025-12-06T15:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.600856 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"393f3b4b-ef95-4f70-af75-9fa123e0d52b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3d7272d34a92fa558ed95d304ebaf3f6f0e79e5d60538187542d201250d93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e297ada3f58d22f81f4014c0d47f04b0fe2e68c1e032e9b17e85907f27f9a6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e297ada3f58d22f81f4014c0d47f04b0fe2e68c1e032e9b17e85907f27f9a6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:18Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.620547 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"012e3f9a-67f1-45f2-8e57-76928bb4d5c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a77b5cb2e8c50b39c3369f156a7800603e7d8bf3d6a544ad0f3a6814361338f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2f1dd12351d9f1ce3ccc9711361cf346d9aa1c4cb5d971020e242c80e557eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8326b53a1a99f7cbd09d55008e9c1b4087993a25ae4b50fafc70e4c7b99138d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9471d3db940ee579d5b60cf1f7750c2c265a8f8e549aaa7ed60e82982cf80d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9471d3db940ee579d5b60cf1f7750c2c265a8f8e549aaa7ed60e82982cf80d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:18Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.641490 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:18Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.658966 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:18Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.680239 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7001bdf05c6acf52c9be35bd63ae103390bf04d2a00a41bf08d376dcdc497338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T15:47:09Z\\\",\\\"message\\\":\\\"2025-12-06T15:46:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5f77eacc-42b0-4fed-a916-5e52a3f47837\\\\n2025-12-06T15:46:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5f77eacc-42b0-4fed-a916-5e52a3f47837 to /host/opt/cni/bin/\\\\n2025-12-06T15:46:24Z [verbose] multus-daemon started\\\\n2025-12-06T15:46:24Z [verbose] Readiness Indicator file check\\\\n2025-12-06T15:47:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:47:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:18Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.698440 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.698566 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.698650 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.698741 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.698811 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:18Z","lastTransitionTime":"2025-12-06T15:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.703408 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:18Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.724984 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:18Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.747979 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a379c66da38cd6d16eb0fdc5941e3fd5d4127b7cce84fe2bde55918ef2c99c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:18Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.776060 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f228acd2a1febab7d6991054e11cf509ff83ab26e537db82ab49688144a4a233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:18Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.802030 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.802131 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.802149 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.802173 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.802190 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:18Z","lastTransitionTime":"2025-12-06T15:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.815075 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:18Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.835073 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ab1d9f15bba233c4bbd04ece144a5a8119c168abaeee7d4544269b55ad58ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b14a8cf06c49b6922327f25905466f86e1a0fd0a0b852f751bac6f8cd7db9e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bd4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:18Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.851160 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kzbhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18a94103-5ed7-4e27-a58d-7c989b0f70f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kzbhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:18Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.870035 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:18Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.901583 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b654f2820e7e70a18617e577d1dffa10f2e6e786d8e8a2483493d1e2cbf32e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d6b3583626fb73a7ee10d9ea994c5dc388215022154a5a51777e14a8195cc0c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T15:46:49Z\\\",\\\"message\\\":\\\"ontroller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:46:49Z is after 2025-08-24T17:21:41Z]\\\\nI1206 15:46:49.421493 6320 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"18746a4d-8a63-458a-b7e3-8fb89ff95fc0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-route-controller-manager/route-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b654f2820e7e70a18617e577d1dffa10f2e6e786d8e8a2483493d1e2cbf32e6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T15:47:18Z\\\",\\\"message\\\":\\\"4-c6801b70dcb0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 15:47:18.251613 6661 services_controller.go:443] Built service openshift-image-registry/image-registry LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.93\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:5000, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1206 15:47:18.251617 6661 ovn.go:134] Ensuring zone local for Pod openshift-kube-scheduler/openshift-kube-scheduler-crc in node crc\\\\nI1206 15:47:18.251627 6661 services_controller.go:444] Built service openshift-image-registry/image-registry LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1206 15:47:18.251637 6661 services_controller.go:445] Built service openshift-image-registry/image-registry LB template configs for network=default: []services.lbConfig(nil)\\\\nF1206 15:47:18.249639 6661 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:18Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.905592 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.905637 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.905657 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.905683 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.905702 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:18Z","lastTransitionTime":"2025-12-06T15:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.915684 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:18Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.928729 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:18Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.943078 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:18Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.956731 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:18Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:18 crc kubenswrapper[4813]: I1206 15:47:18.969583 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:18Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.008731 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.008789 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.008803 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.008827 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.008843 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:19Z","lastTransitionTime":"2025-12-06T15:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.113097 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.113462 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.113594 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.113733 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.113928 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:19Z","lastTransitionTime":"2025-12-06T15:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.216894 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.217519 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.217697 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.217871 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.218054 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:19Z","lastTransitionTime":"2025-12-06T15:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.321790 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.321865 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.321893 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.321924 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.321945 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:19Z","lastTransitionTime":"2025-12-06T15:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.425831 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.425897 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.425916 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.425942 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.425959 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:19Z","lastTransitionTime":"2025-12-06T15:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.485900 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.485954 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.486013 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.485900 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:19 crc kubenswrapper[4813]: E1206 15:47:19.486108 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:47:19 crc kubenswrapper[4813]: E1206 15:47:19.486208 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:47:19 crc kubenswrapper[4813]: E1206 15:47:19.486349 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:47:19 crc kubenswrapper[4813]: E1206 15:47:19.486440 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.533750 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.533845 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.533875 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.533915 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.533954 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:19Z","lastTransitionTime":"2025-12-06T15:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.588883 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xdt4d_0e54e31f-e9f8-4dd5-8738-c785df002c93/ovnkube-controller/3.log" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.595511 4813 scope.go:117] "RemoveContainer" containerID="6b654f2820e7e70a18617e577d1dffa10f2e6e786d8e8a2483493d1e2cbf32e6" Dec 06 15:47:19 crc kubenswrapper[4813]: E1206 15:47:19.595985 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xdt4d_openshift-ovn-kubernetes(0e54e31f-e9f8-4dd5-8738-c785df002c93)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.611742 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:19Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.636940 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.637396 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.637532 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.637651 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.637744 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:19Z","lastTransitionTime":"2025-12-06T15:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.646551 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b654f2820e7e70a18617e577d1dffa10f2e6e786d8e8a2483493d1e2cbf32e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b654f2820e7e70a18617e577d1dffa10f2e6e786d8e8a2483493d1e2cbf32e6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T15:47:18Z\\\",\\\"message\\\":\\\"4-c6801b70dcb0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 15:47:18.251613 6661 services_controller.go:443] Built service openshift-image-registry/image-registry LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.93\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:5000, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1206 15:47:18.251617 6661 ovn.go:134] Ensuring zone local for Pod openshift-kube-scheduler/openshift-kube-scheduler-crc in node crc\\\\nI1206 15:47:18.251627 6661 services_controller.go:444] Built service openshift-image-registry/image-registry LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1206 15:47:18.251637 6661 services_controller.go:445] Built service openshift-image-registry/image-registry LB template configs for network=default: []services.lbConfig(nil)\\\\nF1206 15:47:18.249639 6661 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:47:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xdt4d_openshift-ovn-kubernetes(0e54e31f-e9f8-4dd5-8738-c785df002c93)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:19Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.660998 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:19Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.678362 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:19Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.697339 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:19Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.713720 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:19Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.727293 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:19Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.748556 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.748616 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.748627 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.748648 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.748659 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:19Z","lastTransitionTime":"2025-12-06T15:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.749869 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"393f3b4b-ef95-4f70-af75-9fa123e0d52b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3d7272d34a92fa558ed95d304ebaf3f6f0e79e5d60538187542d201250d93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e297ada3f58d22f81f4014c0d47f04b0fe2e68c1e032e9b17e85907f27f9a6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e297ada3f58d22f81f4014c0d47f04b0fe2e68c1e032e9b17e85907f27f9a6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:19Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.766597 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"012e3f9a-67f1-45f2-8e57-76928bb4d5c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a77b5cb2e8c50b39c3369f156a7800603e7d8bf3d6a544ad0f3a6814361338f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2f1dd12351d9f1ce3ccc9711361cf346d9aa1c4cb5d971020e242c80e557eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8326b53a1a99f7cbd09d55008e9c1b4087993a25ae4b50fafc70e4c7b99138d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9471d3db940ee579d5b60cf1f7750c2c265a8f8e549aaa7ed60e82982cf80d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9471d3db940ee579d5b60cf1f7750c2c265a8f8e549aaa7ed60e82982cf80d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:19Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.779597 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:19Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.791193 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:19Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.804572 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7001bdf05c6acf52c9be35bd63ae103390bf04d2a00a41bf08d376dcdc497338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T15:47:09Z\\\",\\\"message\\\":\\\"2025-12-06T15:46:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5f77eacc-42b0-4fed-a916-5e52a3f47837\\\\n2025-12-06T15:46:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5f77eacc-42b0-4fed-a916-5e52a3f47837 to /host/opt/cni/bin/\\\\n2025-12-06T15:46:24Z [verbose] multus-daemon started\\\\n2025-12-06T15:46:24Z [verbose] Readiness Indicator file check\\\\n2025-12-06T15:47:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:47:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:19Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.820001 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:19Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.832681 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:19Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.845116 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a379c66da38cd6d16eb0fdc5941e3fd5d4127b7cce84fe2bde55918ef2c99c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:19Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.850689 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.850727 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.850739 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.850754 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.850765 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:19Z","lastTransitionTime":"2025-12-06T15:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.860481 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f228acd2a1febab7d6991054e11cf509ff83ab26e537db82ab49688144a4a233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:19Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.897282 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:19Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.910764 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ab1d9f15bba233c4bbd04ece144a5a8119c168abaeee7d4544269b55ad58ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b14a8cf06c49b6922327f25905466f86e1a0fd0a0b852f751bac6f8cd7db9e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bd4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:19Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.922356 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kzbhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18a94103-5ed7-4e27-a58d-7c989b0f70f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kzbhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:19Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.953427 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.953459 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.953467 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.953480 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:19 crc kubenswrapper[4813]: I1206 15:47:19.953489 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:19Z","lastTransitionTime":"2025-12-06T15:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.056460 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.056502 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.056513 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.056528 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.056538 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:20Z","lastTransitionTime":"2025-12-06T15:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.159980 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.160051 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.160076 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.160101 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.160121 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:20Z","lastTransitionTime":"2025-12-06T15:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.263651 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.263707 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.263727 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.263748 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.263763 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:20Z","lastTransitionTime":"2025-12-06T15:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.368116 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.368191 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.368212 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.368305 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.368327 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:20Z","lastTransitionTime":"2025-12-06T15:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.470994 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.471078 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.471103 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.471129 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.471147 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:20Z","lastTransitionTime":"2025-12-06T15:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.506566 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"012e3f9a-67f1-45f2-8e57-76928bb4d5c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a77b5cb2e8c50b39c3369f156a7800603e7d8bf3d6a544ad0f3a6814361338f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2f1dd12351d9f1ce3ccc9711361cf346d9aa1c4cb5d971020e242c80e557eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8326b53a1a99f7cbd09d55008e9c1b4087993a25ae4b50fafc70e4c7b99138d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9471d3db940ee579d5b60cf1f7750c2c265a8f8e549aaa7ed60e82982cf80d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9471d3db940ee579d5b60cf1f7750c2c265a8f8e549aaa7ed60e82982cf80d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:20Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.525884 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:20Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.544905 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:20Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.560339 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"393f3b4b-ef95-4f70-af75-9fa123e0d52b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3d7272d34a92fa558ed95d304ebaf3f6f0e79e5d60538187542d201250d93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e297ada3f58d22f81f4014c0d47f04b0fe2e68c1e032e9b17e85907f27f9a6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e297ada3f58d22f81f4014c0d47f04b0fe2e68c1e032e9b17e85907f27f9a6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:20Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.574442 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.574511 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.574531 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.574564 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.574586 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:20Z","lastTransitionTime":"2025-12-06T15:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.589584 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:20Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.608914 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a379c66da38cd6d16eb0fdc5941e3fd5d4127b7cce84fe2bde55918ef2c99c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:20Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.642727 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f228acd2a1febab7d6991054e11cf509ff83ab26e537db82ab49688144a4a233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:20Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.665160 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7001bdf05c6acf52c9be35bd63ae103390bf04d2a00a41bf08d376dcdc497338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T15:47:09Z\\\",\\\"message\\\":\\\"2025-12-06T15:46:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5f77eacc-42b0-4fed-a916-5e52a3f47837\\\\n2025-12-06T15:46:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5f77eacc-42b0-4fed-a916-5e52a3f47837 to /host/opt/cni/bin/\\\\n2025-12-06T15:46:24Z [verbose] multus-daemon started\\\\n2025-12-06T15:46:24Z [verbose] Readiness Indicator file check\\\\n2025-12-06T15:47:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:47:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:20Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.677771 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.677846 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.677873 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.677902 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.677924 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:20Z","lastTransitionTime":"2025-12-06T15:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.684822 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:20Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.697872 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ab1d9f15bba233c4bbd04ece144a5a8119c168abaeee7d4544269b55ad58ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b14a8cf06c49b6922327f25905466f86e1a0fd0a0b852f751bac6f8cd7db9e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bd4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:20Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.709768 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kzbhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18a94103-5ed7-4e27-a58d-7c989b0f70f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kzbhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:20Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.734334 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:20Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.752390 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:20Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.767671 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:20Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.779643 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.779669 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.779680 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.779696 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.779707 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:20Z","lastTransitionTime":"2025-12-06T15:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.780957 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:20Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.791568 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:20Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.812725 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b654f2820e7e70a18617e577d1dffa10f2e6e786d8e8a2483493d1e2cbf32e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b654f2820e7e70a18617e577d1dffa10f2e6e786d8e8a2483493d1e2cbf32e6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T15:47:18Z\\\",\\\"message\\\":\\\"4-c6801b70dcb0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 15:47:18.251613 6661 services_controller.go:443] Built service openshift-image-registry/image-registry LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.93\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:5000, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1206 15:47:18.251617 6661 ovn.go:134] Ensuring zone local for Pod openshift-kube-scheduler/openshift-kube-scheduler-crc in node crc\\\\nI1206 15:47:18.251627 6661 services_controller.go:444] Built service openshift-image-registry/image-registry LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1206 15:47:18.251637 6661 services_controller.go:445] Built service openshift-image-registry/image-registry LB template configs for network=default: []services.lbConfig(nil)\\\\nF1206 15:47:18.249639 6661 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:47:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xdt4d_openshift-ovn-kubernetes(0e54e31f-e9f8-4dd5-8738-c785df002c93)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:20Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.824417 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:20Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.839059 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:20Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.883339 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.883400 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.883423 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.883455 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.883477 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:20Z","lastTransitionTime":"2025-12-06T15:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.986405 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.986446 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.986458 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.986474 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:20 crc kubenswrapper[4813]: I1206 15:47:20.986486 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:20Z","lastTransitionTime":"2025-12-06T15:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.089618 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.089675 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.089692 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.089716 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.089735 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:21Z","lastTransitionTime":"2025-12-06T15:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.192616 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.192653 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.192665 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.192684 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.192697 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:21Z","lastTransitionTime":"2025-12-06T15:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.295823 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.296217 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.296482 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.296657 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.296875 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:21Z","lastTransitionTime":"2025-12-06T15:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.400459 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.400518 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.400540 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.400572 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.400595 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:21Z","lastTransitionTime":"2025-12-06T15:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.486603 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.486689 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.486638 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.486826 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:21 crc kubenswrapper[4813]: E1206 15:47:21.486987 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:47:21 crc kubenswrapper[4813]: E1206 15:47:21.487360 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:47:21 crc kubenswrapper[4813]: E1206 15:47:21.487698 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:47:21 crc kubenswrapper[4813]: E1206 15:47:21.487760 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.504333 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.504580 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.504779 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.504945 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.505090 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:21Z","lastTransitionTime":"2025-12-06T15:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.608313 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.608722 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.608867 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.609109 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.609390 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:21Z","lastTransitionTime":"2025-12-06T15:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.712659 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.712720 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.712736 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.712759 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.712776 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:21Z","lastTransitionTime":"2025-12-06T15:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.815591 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.815651 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.815668 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.815693 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.815711 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:21Z","lastTransitionTime":"2025-12-06T15:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.919064 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.919103 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.919112 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.919130 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:21 crc kubenswrapper[4813]: I1206 15:47:21.919141 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:21Z","lastTransitionTime":"2025-12-06T15:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.021847 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.022127 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.022253 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.022456 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.022625 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:22Z","lastTransitionTime":"2025-12-06T15:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.125048 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.125100 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.125123 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.125144 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.125158 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:22Z","lastTransitionTime":"2025-12-06T15:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.228072 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.228117 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.228128 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.228146 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.228158 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:22Z","lastTransitionTime":"2025-12-06T15:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.332452 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.332531 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.332551 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.332574 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.332591 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:22Z","lastTransitionTime":"2025-12-06T15:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.435868 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.435933 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.435951 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.435975 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.435993 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:22Z","lastTransitionTime":"2025-12-06T15:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.539025 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.539124 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.539143 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.539170 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.539188 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:22Z","lastTransitionTime":"2025-12-06T15:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.642688 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.642754 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.642774 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.642797 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.642816 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:22Z","lastTransitionTime":"2025-12-06T15:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.746678 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.746740 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.746759 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.746786 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.746805 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:22Z","lastTransitionTime":"2025-12-06T15:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.849538 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.849597 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.849609 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.849628 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.849640 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:22Z","lastTransitionTime":"2025-12-06T15:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.953081 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.953159 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.953177 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.953204 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:22 crc kubenswrapper[4813]: I1206 15:47:22.953228 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:22Z","lastTransitionTime":"2025-12-06T15:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.057475 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.057553 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.057570 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.057597 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.057616 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:23Z","lastTransitionTime":"2025-12-06T15:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.162799 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.162852 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.162868 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.162891 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.162908 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:23Z","lastTransitionTime":"2025-12-06T15:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.265087 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.265143 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.265154 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.265172 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.265185 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:23Z","lastTransitionTime":"2025-12-06T15:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.329865 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.330020 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:23 crc kubenswrapper[4813]: E1206 15:47:23.330051 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:27.33001228 +0000 UTC m=+147.220891886 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.330123 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:23 crc kubenswrapper[4813]: E1206 15:47:23.330134 4813 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 15:47:23 crc kubenswrapper[4813]: E1206 15:47:23.330201 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 15:48:27.330184015 +0000 UTC m=+147.221063601 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.330234 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:23 crc kubenswrapper[4813]: E1206 15:47:23.330293 4813 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 15:47:23 crc kubenswrapper[4813]: E1206 15:47:23.330349 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 15:48:27.330335119 +0000 UTC m=+147.221214735 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 15:47:23 crc kubenswrapper[4813]: E1206 15:47:23.330368 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 15:47:23 crc kubenswrapper[4813]: E1206 15:47:23.330385 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 15:47:23 crc kubenswrapper[4813]: E1206 15:47:23.330399 4813 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:47:23 crc kubenswrapper[4813]: E1206 15:47:23.330443 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 15:48:27.330433392 +0000 UTC m=+147.221312978 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.368617 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.368677 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.368692 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.368714 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.368730 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:23Z","lastTransitionTime":"2025-12-06T15:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.472159 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.472239 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.472304 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.472339 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.472363 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:23Z","lastTransitionTime":"2025-12-06T15:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.486411 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.486519 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:23 crc kubenswrapper[4813]: E1206 15:47:23.486600 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.486701 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.486754 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:23 crc kubenswrapper[4813]: E1206 15:47:23.486886 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:47:23 crc kubenswrapper[4813]: E1206 15:47:23.487019 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:47:23 crc kubenswrapper[4813]: E1206 15:47:23.487100 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.575475 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.575522 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.575534 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.575550 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.575562 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:23Z","lastTransitionTime":"2025-12-06T15:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.633726 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:23 crc kubenswrapper[4813]: E1206 15:47:23.633943 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 15:47:23 crc kubenswrapper[4813]: E1206 15:47:23.633987 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 15:47:23 crc kubenswrapper[4813]: E1206 15:47:23.634006 4813 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:47:23 crc kubenswrapper[4813]: E1206 15:47:23.634088 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 15:48:27.634064239 +0000 UTC m=+147.524943845 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.677826 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.677888 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.677902 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.677926 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.677941 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:23Z","lastTransitionTime":"2025-12-06T15:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.781105 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.781144 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.781157 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.781171 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.781181 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:23Z","lastTransitionTime":"2025-12-06T15:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.884470 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.884521 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.884540 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.884568 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.884586 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:23Z","lastTransitionTime":"2025-12-06T15:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.987171 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.987249 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.987295 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.987327 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:23 crc kubenswrapper[4813]: I1206 15:47:23.987347 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:23Z","lastTransitionTime":"2025-12-06T15:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.090229 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.090296 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.090308 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.090325 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.090336 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:24Z","lastTransitionTime":"2025-12-06T15:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.193605 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.193654 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.193675 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.193698 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.193714 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:24Z","lastTransitionTime":"2025-12-06T15:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.297172 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.297231 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.297243 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.297278 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.297294 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:24Z","lastTransitionTime":"2025-12-06T15:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.400708 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.400783 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.400804 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.400828 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.400844 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:24Z","lastTransitionTime":"2025-12-06T15:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.503165 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.503222 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.503240 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.503291 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.503311 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:24Z","lastTransitionTime":"2025-12-06T15:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.605320 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.605539 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.605547 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.605560 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.605569 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:24Z","lastTransitionTime":"2025-12-06T15:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.708602 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.708663 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.708683 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.708706 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.708724 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:24Z","lastTransitionTime":"2025-12-06T15:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.811840 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.811892 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.811908 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.811938 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.811956 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:24Z","lastTransitionTime":"2025-12-06T15:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.915834 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.915892 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.915910 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.915937 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:24 crc kubenswrapper[4813]: I1206 15:47:24.915956 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:24Z","lastTransitionTime":"2025-12-06T15:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.019119 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.019189 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.019212 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.019240 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.019293 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:25Z","lastTransitionTime":"2025-12-06T15:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.122352 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.122419 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.122439 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.122464 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.122483 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:25Z","lastTransitionTime":"2025-12-06T15:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.224948 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.225040 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.225058 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.225083 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.225099 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:25Z","lastTransitionTime":"2025-12-06T15:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.329056 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.329130 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.329155 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.329185 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.329210 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:25Z","lastTransitionTime":"2025-12-06T15:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.432661 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.432732 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.432753 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.432778 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.432797 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:25Z","lastTransitionTime":"2025-12-06T15:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.486785 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.486834 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.486872 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.486809 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:25 crc kubenswrapper[4813]: E1206 15:47:25.486958 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:47:25 crc kubenswrapper[4813]: E1206 15:47:25.487120 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:47:25 crc kubenswrapper[4813]: E1206 15:47:25.487205 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:47:25 crc kubenswrapper[4813]: E1206 15:47:25.487303 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.536144 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.536220 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.536236 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.536308 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.536329 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:25Z","lastTransitionTime":"2025-12-06T15:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.639492 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.639546 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.639565 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.639587 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.639603 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:25Z","lastTransitionTime":"2025-12-06T15:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.742022 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.742097 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.742122 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.742154 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.742179 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:25Z","lastTransitionTime":"2025-12-06T15:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.845603 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.845667 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.845688 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.845712 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.845730 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:25Z","lastTransitionTime":"2025-12-06T15:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.948564 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.948617 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.948632 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.948654 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:25 crc kubenswrapper[4813]: I1206 15:47:25.948668 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:25Z","lastTransitionTime":"2025-12-06T15:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.051655 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.051702 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.051714 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.051733 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.051746 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:26Z","lastTransitionTime":"2025-12-06T15:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.154822 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.154873 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.154889 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.154911 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.154928 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:26Z","lastTransitionTime":"2025-12-06T15:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.258595 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.258650 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.258665 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.258714 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.258730 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:26Z","lastTransitionTime":"2025-12-06T15:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.362754 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.362819 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.362843 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.362875 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.362901 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:26Z","lastTransitionTime":"2025-12-06T15:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.466033 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.466093 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.466111 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.466136 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.466155 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:26Z","lastTransitionTime":"2025-12-06T15:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.568580 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.568627 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.568640 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.568658 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.568671 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:26Z","lastTransitionTime":"2025-12-06T15:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.671798 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.671872 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.671892 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.671914 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.671930 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:26Z","lastTransitionTime":"2025-12-06T15:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.773970 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.774032 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.774050 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.774074 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.774093 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:26Z","lastTransitionTime":"2025-12-06T15:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.877412 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.877502 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.877527 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.877559 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.877581 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:26Z","lastTransitionTime":"2025-12-06T15:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.980691 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.980758 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.980775 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.980799 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.980816 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:26Z","lastTransitionTime":"2025-12-06T15:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.999451 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.999510 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.999527 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.999550 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:26 crc kubenswrapper[4813]: I1206 15:47:26.999567 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:26Z","lastTransitionTime":"2025-12-06T15:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:27 crc kubenswrapper[4813]: E1206 15:47:27.018409 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:27Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.023335 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.023394 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.023416 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.023446 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.023469 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:27Z","lastTransitionTime":"2025-12-06T15:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:27 crc kubenswrapper[4813]: E1206 15:47:27.042362 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:27Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.048055 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.048234 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.048264 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.048347 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.048375 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:27Z","lastTransitionTime":"2025-12-06T15:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:27 crc kubenswrapper[4813]: E1206 15:47:27.072501 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:27Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.081213 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.081252 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.081276 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.081291 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.081305 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:27Z","lastTransitionTime":"2025-12-06T15:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:27 crc kubenswrapper[4813]: E1206 15:47:27.097877 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:27Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.101808 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.101832 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.101841 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.101855 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.101865 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:27Z","lastTransitionTime":"2025-12-06T15:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:27 crc kubenswrapper[4813]: E1206 15:47:27.114920 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:27Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:27 crc kubenswrapper[4813]: E1206 15:47:27.115046 4813 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.116346 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.116380 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.116392 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.116407 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.116427 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:27Z","lastTransitionTime":"2025-12-06T15:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.218923 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.218970 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.218983 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.219000 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.219012 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:27Z","lastTransitionTime":"2025-12-06T15:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.322002 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.322078 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.322100 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.322127 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.322148 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:27Z","lastTransitionTime":"2025-12-06T15:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.425053 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.425129 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.425148 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.425171 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.425187 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:27Z","lastTransitionTime":"2025-12-06T15:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.485768 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.485838 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.485872 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.485884 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:27 crc kubenswrapper[4813]: E1206 15:47:27.486086 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:47:27 crc kubenswrapper[4813]: E1206 15:47:27.486237 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:47:27 crc kubenswrapper[4813]: E1206 15:47:27.486393 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:47:27 crc kubenswrapper[4813]: E1206 15:47:27.486464 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.528292 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.528356 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.528379 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.528407 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.528428 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:27Z","lastTransitionTime":"2025-12-06T15:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.630730 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.630762 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.630771 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.630794 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.630806 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:27Z","lastTransitionTime":"2025-12-06T15:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.733452 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.733485 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.733498 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.733515 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.733526 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:27Z","lastTransitionTime":"2025-12-06T15:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.836872 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.836947 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.836970 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.836999 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.837020 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:27Z","lastTransitionTime":"2025-12-06T15:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.939791 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.939849 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.939861 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.939877 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:27 crc kubenswrapper[4813]: I1206 15:47:27.939919 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:27Z","lastTransitionTime":"2025-12-06T15:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.042025 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.042084 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.042103 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.042128 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.042155 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:28Z","lastTransitionTime":"2025-12-06T15:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.145062 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.145129 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.145147 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.145171 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.145188 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:28Z","lastTransitionTime":"2025-12-06T15:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.249167 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.249572 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.249841 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.250051 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.250229 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:28Z","lastTransitionTime":"2025-12-06T15:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.353942 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.354002 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.354021 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.354048 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.354069 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:28Z","lastTransitionTime":"2025-12-06T15:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.456479 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.456526 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.456538 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.456556 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.456568 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:28Z","lastTransitionTime":"2025-12-06T15:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.559470 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.560431 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.560629 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.560855 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.561034 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:28Z","lastTransitionTime":"2025-12-06T15:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.664046 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.664086 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.664097 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.664112 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.664123 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:28Z","lastTransitionTime":"2025-12-06T15:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.767800 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.768193 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.768392 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.768522 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.768646 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:28Z","lastTransitionTime":"2025-12-06T15:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.871630 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.871693 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.871711 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.871735 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.871755 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:28Z","lastTransitionTime":"2025-12-06T15:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.974427 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.974489 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.974508 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.974534 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:28 crc kubenswrapper[4813]: I1206 15:47:28.974553 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:28Z","lastTransitionTime":"2025-12-06T15:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.077188 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.077288 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.077299 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.077319 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.077330 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:29Z","lastTransitionTime":"2025-12-06T15:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.180975 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.181015 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.181024 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.181039 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.181050 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:29Z","lastTransitionTime":"2025-12-06T15:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.284316 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.284369 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.284382 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.284403 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.284415 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:29Z","lastTransitionTime":"2025-12-06T15:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.394822 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.394888 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.394924 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.394954 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.394974 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:29Z","lastTransitionTime":"2025-12-06T15:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.485799 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.485940 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:29 crc kubenswrapper[4813]: E1206 15:47:29.485967 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.486001 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:29 crc kubenswrapper[4813]: E1206 15:47:29.486152 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.486191 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:29 crc kubenswrapper[4813]: E1206 15:47:29.486290 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:47:29 crc kubenswrapper[4813]: E1206 15:47:29.486348 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.497120 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.497165 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.497179 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.497194 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.497207 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:29Z","lastTransitionTime":"2025-12-06T15:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.600246 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.600409 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.600431 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.600457 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.600473 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:29Z","lastTransitionTime":"2025-12-06T15:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.703784 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.703876 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.703902 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.703939 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.703969 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:29Z","lastTransitionTime":"2025-12-06T15:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.806224 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.806294 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.806308 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.806331 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.806345 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:29Z","lastTransitionTime":"2025-12-06T15:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.908544 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.908583 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.908591 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.908607 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:29 crc kubenswrapper[4813]: I1206 15:47:29.908617 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:29Z","lastTransitionTime":"2025-12-06T15:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.012014 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.012053 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.012066 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.012082 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.012093 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:30Z","lastTransitionTime":"2025-12-06T15:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.114477 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.114509 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.114517 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.114529 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.114538 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:30Z","lastTransitionTime":"2025-12-06T15:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.217630 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.217674 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.217687 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.217705 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.217716 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:30Z","lastTransitionTime":"2025-12-06T15:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.321034 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.321192 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.321226 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.321326 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.321355 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:30Z","lastTransitionTime":"2025-12-06T15:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.424054 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.424112 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.424120 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.424136 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.424144 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:30Z","lastTransitionTime":"2025-12-06T15:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.508224 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.528189 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.528807 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.528989 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.529140 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.529329 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.529490 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:30Z","lastTransitionTime":"2025-12-06T15:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.542103 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.557425 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.587239 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b654f2820e7e70a18617e577d1dffa10f2e6e786d8e8a2483493d1e2cbf32e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b654f2820e7e70a18617e577d1dffa10f2e6e786d8e8a2483493d1e2cbf32e6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T15:47:18Z\\\",\\\"message\\\":\\\"4-c6801b70dcb0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 15:47:18.251613 6661 services_controller.go:443] Built service openshift-image-registry/image-registry LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.93\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:5000, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1206 15:47:18.251617 6661 ovn.go:134] Ensuring zone local for Pod openshift-kube-scheduler/openshift-kube-scheduler-crc in node crc\\\\nI1206 15:47:18.251627 6661 services_controller.go:444] Built service openshift-image-registry/image-registry LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1206 15:47:18.251637 6661 services_controller.go:445] Built service openshift-image-registry/image-registry LB template configs for network=default: []services.lbConfig(nil)\\\\nF1206 15:47:18.249639 6661 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:47:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xdt4d_openshift-ovn-kubernetes(0e54e31f-e9f8-4dd5-8738-c785df002c93)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.604714 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.622340 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.632759 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.632795 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.632806 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.632821 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.632834 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:30Z","lastTransitionTime":"2025-12-06T15:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.639032 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"012e3f9a-67f1-45f2-8e57-76928bb4d5c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a77b5cb2e8c50b39c3369f156a7800603e7d8bf3d6a544ad0f3a6814361338f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2f1dd12351d9f1ce3ccc9711361cf346d9aa1c4cb5d971020e242c80e557eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8326b53a1a99f7cbd09d55008e9c1b4087993a25ae4b50fafc70e4c7b99138d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9471d3db940ee579d5b60cf1f7750c2c265a8f8e549aaa7ed60e82982cf80d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9471d3db940ee579d5b60cf1f7750c2c265a8f8e549aaa7ed60e82982cf80d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.654902 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.667071 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.679925 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"393f3b4b-ef95-4f70-af75-9fa123e0d52b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3d7272d34a92fa558ed95d304ebaf3f6f0e79e5d60538187542d201250d93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e297ada3f58d22f81f4014c0d47f04b0fe2e68c1e032e9b17e85907f27f9a6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e297ada3f58d22f81f4014c0d47f04b0fe2e68c1e032e9b17e85907f27f9a6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.697868 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.715771 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a379c66da38cd6d16eb0fdc5941e3fd5d4127b7cce84fe2bde55918ef2c99c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.733189 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f228acd2a1febab7d6991054e11cf509ff83ab26e537db82ab49688144a4a233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.736345 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.736393 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.736408 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.736429 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.736443 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:30Z","lastTransitionTime":"2025-12-06T15:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.749058 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7001bdf05c6acf52c9be35bd63ae103390bf04d2a00a41bf08d376dcdc497338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T15:47:09Z\\\",\\\"message\\\":\\\"2025-12-06T15:46:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5f77eacc-42b0-4fed-a916-5e52a3f47837\\\\n2025-12-06T15:46:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5f77eacc-42b0-4fed-a916-5e52a3f47837 to /host/opt/cni/bin/\\\\n2025-12-06T15:46:24Z [verbose] multus-daemon started\\\\n2025-12-06T15:46:24Z [verbose] Readiness Indicator file check\\\\n2025-12-06T15:47:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:47:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.765782 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.775915 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ab1d9f15bba233c4bbd04ece144a5a8119c168abaeee7d4544269b55ad58ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b14a8cf06c49b6922327f25905466f86e1a0fd0a0b852f751bac6f8cd7db9e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bd4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.785276 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kzbhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18a94103-5ed7-4e27-a58d-7c989b0f70f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kzbhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.804489 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:30Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.839520 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.839812 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.839947 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.840089 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.840215 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:30Z","lastTransitionTime":"2025-12-06T15:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.943872 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.943913 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.943925 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.943941 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:30 crc kubenswrapper[4813]: I1206 15:47:30.943951 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:30Z","lastTransitionTime":"2025-12-06T15:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.045599 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.045656 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.045679 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.045707 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.045727 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:31Z","lastTransitionTime":"2025-12-06T15:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.148857 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.148896 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.148907 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.148924 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.148936 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:31Z","lastTransitionTime":"2025-12-06T15:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.251917 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.251985 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.252010 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.252033 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.252046 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:31Z","lastTransitionTime":"2025-12-06T15:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.355752 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.355890 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.355912 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.355936 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.355954 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:31Z","lastTransitionTime":"2025-12-06T15:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.460335 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.460439 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.460492 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.460517 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.460532 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:31Z","lastTransitionTime":"2025-12-06T15:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.486865 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.486943 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.487015 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:31 crc kubenswrapper[4813]: E1206 15:47:31.487729 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:47:31 crc kubenswrapper[4813]: E1206 15:47:31.487903 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:47:31 crc kubenswrapper[4813]: E1206 15:47:31.488178 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.488305 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:31 crc kubenswrapper[4813]: E1206 15:47:31.488464 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.564749 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.564814 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.564833 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.564862 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.564882 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:31Z","lastTransitionTime":"2025-12-06T15:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.667900 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.668359 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.668429 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.668505 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.668614 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:31Z","lastTransitionTime":"2025-12-06T15:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.805454 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.805518 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.805535 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.805559 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.805577 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:31Z","lastTransitionTime":"2025-12-06T15:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.908489 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.908547 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.908565 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.908590 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:31 crc kubenswrapper[4813]: I1206 15:47:31.908608 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:31Z","lastTransitionTime":"2025-12-06T15:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.012039 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.012455 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.012678 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.012903 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.013122 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:32Z","lastTransitionTime":"2025-12-06T15:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.117043 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.117094 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.117115 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.117137 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.117154 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:32Z","lastTransitionTime":"2025-12-06T15:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.220820 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.220888 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.220909 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.220937 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.220957 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:32Z","lastTransitionTime":"2025-12-06T15:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.324691 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.325055 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.325436 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.325851 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.326212 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:32Z","lastTransitionTime":"2025-12-06T15:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.431008 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.431085 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.431106 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.431135 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.431158 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:32Z","lastTransitionTime":"2025-12-06T15:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.534359 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.534418 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.534435 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.534456 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.534474 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:32Z","lastTransitionTime":"2025-12-06T15:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.637258 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.637353 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.637377 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.637404 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.637424 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:32Z","lastTransitionTime":"2025-12-06T15:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.740992 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.741058 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.741080 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.741111 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.741132 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:32Z","lastTransitionTime":"2025-12-06T15:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.845319 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.845716 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.845874 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.846036 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.846185 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:32Z","lastTransitionTime":"2025-12-06T15:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.949716 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.949835 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.949862 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.949891 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:32 crc kubenswrapper[4813]: I1206 15:47:32.949914 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:32Z","lastTransitionTime":"2025-12-06T15:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.052662 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.052726 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.052743 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.052767 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.052784 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:33Z","lastTransitionTime":"2025-12-06T15:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.155675 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.155748 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.155773 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.155802 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.155828 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:33Z","lastTransitionTime":"2025-12-06T15:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.259472 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.259709 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.259771 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.259803 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.259827 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:33Z","lastTransitionTime":"2025-12-06T15:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.362710 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.362769 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.362785 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.362812 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.362833 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:33Z","lastTransitionTime":"2025-12-06T15:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.465609 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.465681 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.465701 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.465728 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.465748 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:33Z","lastTransitionTime":"2025-12-06T15:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.485895 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.485961 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.485970 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.486036 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:33 crc kubenswrapper[4813]: E1206 15:47:33.486239 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:47:33 crc kubenswrapper[4813]: E1206 15:47:33.486479 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:47:33 crc kubenswrapper[4813]: E1206 15:47:33.487122 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:47:33 crc kubenswrapper[4813]: E1206 15:47:33.487219 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.487669 4813 scope.go:117] "RemoveContainer" containerID="6b654f2820e7e70a18617e577d1dffa10f2e6e786d8e8a2483493d1e2cbf32e6" Dec 06 15:47:33 crc kubenswrapper[4813]: E1206 15:47:33.487990 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xdt4d_openshift-ovn-kubernetes(0e54e31f-e9f8-4dd5-8738-c785df002c93)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.568753 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.568812 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.568824 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.568841 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.568855 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:33Z","lastTransitionTime":"2025-12-06T15:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.671320 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.671363 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.671375 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.671398 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.671411 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:33Z","lastTransitionTime":"2025-12-06T15:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.773901 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.773970 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.773996 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.774039 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.774063 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:33Z","lastTransitionTime":"2025-12-06T15:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.877830 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.877907 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.877929 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.877961 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.877984 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:33Z","lastTransitionTime":"2025-12-06T15:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.981695 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.981781 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.981806 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.981840 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:33 crc kubenswrapper[4813]: I1206 15:47:33.981866 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:33Z","lastTransitionTime":"2025-12-06T15:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.088321 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.088407 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.088429 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.088463 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.088491 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:34Z","lastTransitionTime":"2025-12-06T15:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.192174 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.192521 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.192640 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.192746 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.192848 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:34Z","lastTransitionTime":"2025-12-06T15:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.295449 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.295510 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.295529 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.295554 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.295571 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:34Z","lastTransitionTime":"2025-12-06T15:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.398865 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.398919 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.398937 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.398961 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.398977 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:34Z","lastTransitionTime":"2025-12-06T15:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.501943 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.502001 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.502019 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.502042 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.502062 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:34Z","lastTransitionTime":"2025-12-06T15:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.605320 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.605472 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.605544 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.605569 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.605587 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:34Z","lastTransitionTime":"2025-12-06T15:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.708868 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.708937 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.708955 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.708980 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.708997 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:34Z","lastTransitionTime":"2025-12-06T15:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.812836 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.812895 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.812912 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.812935 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.812952 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:34Z","lastTransitionTime":"2025-12-06T15:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.915581 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.915917 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.915942 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.915972 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:34 crc kubenswrapper[4813]: I1206 15:47:34.915993 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:34Z","lastTransitionTime":"2025-12-06T15:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.019006 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.019107 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.019128 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.019156 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.019177 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:35Z","lastTransitionTime":"2025-12-06T15:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.121881 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.121913 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.121929 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.121947 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.121957 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:35Z","lastTransitionTime":"2025-12-06T15:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.224607 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.225119 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.225286 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.225441 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.225571 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:35Z","lastTransitionTime":"2025-12-06T15:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.328065 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.328951 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.329126 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.329254 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.329431 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:35Z","lastTransitionTime":"2025-12-06T15:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.432336 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.432402 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.432421 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.432451 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.432471 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:35Z","lastTransitionTime":"2025-12-06T15:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.486800 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.486829 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.486833 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.486918 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:35 crc kubenswrapper[4813]: E1206 15:47:35.486989 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:47:35 crc kubenswrapper[4813]: E1206 15:47:35.486935 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:47:35 crc kubenswrapper[4813]: E1206 15:47:35.487172 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:47:35 crc kubenswrapper[4813]: E1206 15:47:35.487230 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.535076 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.535127 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.535143 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.535164 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.535180 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:35Z","lastTransitionTime":"2025-12-06T15:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.637736 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.637788 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.637807 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.637829 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.637846 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:35Z","lastTransitionTime":"2025-12-06T15:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.743244 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.743398 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.743418 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.743443 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.743471 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:35Z","lastTransitionTime":"2025-12-06T15:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.846442 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.846492 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.846507 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.846530 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.846546 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:35Z","lastTransitionTime":"2025-12-06T15:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.949895 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.950255 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.950598 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.950865 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:35 crc kubenswrapper[4813]: I1206 15:47:35.951072 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:35Z","lastTransitionTime":"2025-12-06T15:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.054179 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.054238 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.054256 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.054314 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.054333 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:36Z","lastTransitionTime":"2025-12-06T15:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.157430 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.157738 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.157834 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.157938 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.158029 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:36Z","lastTransitionTime":"2025-12-06T15:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.260847 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.261116 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.261195 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.261306 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.261385 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:36Z","lastTransitionTime":"2025-12-06T15:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.364890 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.364946 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.364963 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.364988 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.365006 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:36Z","lastTransitionTime":"2025-12-06T15:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.468368 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.468437 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.468454 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.468479 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.468496 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:36Z","lastTransitionTime":"2025-12-06T15:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.570629 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.570692 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.570708 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.570733 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.570750 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:36Z","lastTransitionTime":"2025-12-06T15:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.673209 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.673637 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.673940 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.674438 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.674575 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:36Z","lastTransitionTime":"2025-12-06T15:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.778110 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.778440 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.778610 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.778782 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.778906 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:36Z","lastTransitionTime":"2025-12-06T15:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.882354 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.882776 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.882939 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.883089 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.883206 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:36Z","lastTransitionTime":"2025-12-06T15:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.986024 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.986070 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.986086 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.986109 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:36 crc kubenswrapper[4813]: I1206 15:47:36.986126 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:36Z","lastTransitionTime":"2025-12-06T15:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.089248 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.089359 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.089378 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.089402 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.089421 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:37Z","lastTransitionTime":"2025-12-06T15:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.192665 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.192724 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.192743 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.192767 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.192785 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:37Z","lastTransitionTime":"2025-12-06T15:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:37 crc kubenswrapper[4813]: E1206 15:47:37.212840 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:37Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.218646 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.218759 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.219158 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.219186 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.219202 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:37Z","lastTransitionTime":"2025-12-06T15:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:37 crc kubenswrapper[4813]: E1206 15:47:37.242755 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:37Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.248339 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.248436 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.248457 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.248483 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.248532 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:37Z","lastTransitionTime":"2025-12-06T15:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:37 crc kubenswrapper[4813]: E1206 15:47:37.269180 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:37Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.274168 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.274242 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.274300 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.274334 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.274359 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:37Z","lastTransitionTime":"2025-12-06T15:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:37 crc kubenswrapper[4813]: E1206 15:47:37.295986 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:37Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.300415 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.300602 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.300699 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.300799 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.300925 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:37Z","lastTransitionTime":"2025-12-06T15:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:37 crc kubenswrapper[4813]: E1206 15:47:37.316305 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1c54b507-6054-49c8-a12e-f79e691f7aaa\\\",\\\"systemUUID\\\":\\\"38943010-e20a-4c38-8e12-0258df15005d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:37Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:37 crc kubenswrapper[4813]: E1206 15:47:37.316470 4813 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.318119 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.318160 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.318176 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.318196 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.318211 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:37Z","lastTransitionTime":"2025-12-06T15:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.420782 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.420837 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.420859 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.420888 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.420911 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:37Z","lastTransitionTime":"2025-12-06T15:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.485763 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.485862 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:37 crc kubenswrapper[4813]: E1206 15:47:37.485963 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.485993 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:37 crc kubenswrapper[4813]: E1206 15:47:37.486160 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.486203 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:37 crc kubenswrapper[4813]: E1206 15:47:37.486556 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:47:37 crc kubenswrapper[4813]: E1206 15:47:37.486660 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.524107 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.524179 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.524200 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.524224 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.524240 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:37Z","lastTransitionTime":"2025-12-06T15:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.627239 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.627336 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.627362 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.627392 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.627416 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:37Z","lastTransitionTime":"2025-12-06T15:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.732441 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.732505 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.732524 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.732548 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.732565 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:37Z","lastTransitionTime":"2025-12-06T15:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.836233 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.836341 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.836367 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.836420 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.836443 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:37Z","lastTransitionTime":"2025-12-06T15:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.941380 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.941431 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.941452 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.941477 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:37 crc kubenswrapper[4813]: I1206 15:47:37.941494 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:37Z","lastTransitionTime":"2025-12-06T15:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.045254 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.045669 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.045846 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.046017 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.046189 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:38Z","lastTransitionTime":"2025-12-06T15:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.149372 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.149429 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.149446 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.149470 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.149490 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:38Z","lastTransitionTime":"2025-12-06T15:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.252490 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.252543 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.252573 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.252595 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.252612 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:38Z","lastTransitionTime":"2025-12-06T15:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.355611 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.355667 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.355693 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.355724 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.355746 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:38Z","lastTransitionTime":"2025-12-06T15:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.458020 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.458095 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.458107 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.458126 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.458139 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:38Z","lastTransitionTime":"2025-12-06T15:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.561576 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.561651 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.561672 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.561698 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.561716 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:38Z","lastTransitionTime":"2025-12-06T15:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.664232 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.664316 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.664334 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.664356 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.664373 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:38Z","lastTransitionTime":"2025-12-06T15:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.767287 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.767340 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.767356 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.767419 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.767437 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:38Z","lastTransitionTime":"2025-12-06T15:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.870130 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.870176 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.870194 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.870217 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.870233 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:38Z","lastTransitionTime":"2025-12-06T15:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.973143 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.973204 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.973221 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.973313 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:38 crc kubenswrapper[4813]: I1206 15:47:38.973333 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:38Z","lastTransitionTime":"2025-12-06T15:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.076799 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.076868 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.076891 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.076918 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.076935 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:39Z","lastTransitionTime":"2025-12-06T15:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.179916 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.179973 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.179990 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.180015 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.180032 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:39Z","lastTransitionTime":"2025-12-06T15:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.282767 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.282828 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.282845 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.282871 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.282896 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:39Z","lastTransitionTime":"2025-12-06T15:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.386662 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.386727 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.386777 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.386969 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.386986 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:39Z","lastTransitionTime":"2025-12-06T15:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.411724 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18a94103-5ed7-4e27-a58d-7c989b0f70f3-metrics-certs\") pod \"network-metrics-daemon-kzbhb\" (UID: \"18a94103-5ed7-4e27-a58d-7c989b0f70f3\") " pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:39 crc kubenswrapper[4813]: E1206 15:47:39.411905 4813 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 15:47:39 crc kubenswrapper[4813]: E1206 15:47:39.412249 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18a94103-5ed7-4e27-a58d-7c989b0f70f3-metrics-certs podName:18a94103-5ed7-4e27-a58d-7c989b0f70f3 nodeName:}" failed. No retries permitted until 2025-12-06 15:48:43.412221137 +0000 UTC m=+163.303100753 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/18a94103-5ed7-4e27-a58d-7c989b0f70f3-metrics-certs") pod "network-metrics-daemon-kzbhb" (UID: "18a94103-5ed7-4e27-a58d-7c989b0f70f3") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.494898 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.494898 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.494988 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.496029 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:39 crc kubenswrapper[4813]: E1206 15:47:39.496219 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:47:39 crc kubenswrapper[4813]: E1206 15:47:39.496366 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:47:39 crc kubenswrapper[4813]: E1206 15:47:39.496903 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:47:39 crc kubenswrapper[4813]: E1206 15:47:39.497112 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.497219 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.497298 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.497324 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.497352 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.497371 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:39Z","lastTransitionTime":"2025-12-06T15:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.600403 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.600446 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.600462 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.600484 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.600503 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:39Z","lastTransitionTime":"2025-12-06T15:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.703658 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.703979 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.704060 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.704139 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.704209 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:39Z","lastTransitionTime":"2025-12-06T15:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.807357 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.807753 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.807949 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.808105 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.808244 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:39Z","lastTransitionTime":"2025-12-06T15:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.910756 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.910843 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.910866 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.910895 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:39 crc kubenswrapper[4813]: I1206 15:47:39.910916 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:39Z","lastTransitionTime":"2025-12-06T15:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.014075 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.014122 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.014140 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.014166 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.014188 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:40Z","lastTransitionTime":"2025-12-06T15:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.117544 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.117609 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.117629 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.117654 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.117673 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:40Z","lastTransitionTime":"2025-12-06T15:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.220991 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.221034 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.221049 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.221070 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.221086 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:40Z","lastTransitionTime":"2025-12-06T15:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.324849 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.324900 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.324917 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.324941 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.324958 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:40Z","lastTransitionTime":"2025-12-06T15:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.427781 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.427825 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.427842 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.427864 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.427883 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:40Z","lastTransitionTime":"2025-12-06T15:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.511474 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"378199a2-e738-42f0-9a44-a5722a68076a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\":]:17697\\\\nI1206 15:46:19.846282 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1206 15:46:19.846301 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846316 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1206 15:46:19.846342 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3475984261/tls.crt::/tmp/serving-cert-3475984261/tls.key\\\\\\\"\\\\nI1206 15:46:19.846461 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1206 15:46:19.846594 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846599 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1206 15:46:19.846611 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1206 15:46:19.846615 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1206 15:46:19.846657 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1206 15:46:19.846666 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1206 15:46:19.849095 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849498 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1206 15:46:19.849953 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1206 15:46:19.853713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.530726 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.530828 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.530849 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.530873 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.530930 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:40Z","lastTransitionTime":"2025-12-06T15:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.536648 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3566897ec1fcf71376101eeb099ce709c9a8f33a9917ba87ac728167aca54d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a8b773ab7e24d3cf63c1d7ae59f3fcdb43c0d601d054fcc69d03df6588e0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.552564 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a379c66da38cd6d16eb0fdc5941e3fd5d4127b7cce84fe2bde55918ef2c99c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.576592 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6adb3a02-198b-41b4-8b18-56b0f9527fe7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f228acd2a1febab7d6991054e11cf509ff83ab26e537db82ab49688144a4a233\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edbce273d9881aae1bf4c960ab9c78204f74f90153087de60ddd6711e3ce8b85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://511f480628c624c54064a59a71d324ef3a00e856fa67111e98e13058635de378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a00cfb43d94d0bb44f1e21fccd7523fef51efb822d39f07a7a8fbb6a355eff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b19b65eb8d0ff66209723082b1b0aae401a3fd8dfcc7fa33c20f72610967548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://059255ef68e73d3e050c5a89e56e4d06b4ed89b489030c5d2451c9e91500e903\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8bb8d04406cdf620d38f100ea94d35928a813d25314dae9fd03cee3ce57c1a82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k49xc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t9zmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.601385 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gl495" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856bb45b-525c-4ef0-bf7c-0691cf54b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:47:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7001bdf05c6acf52c9be35bd63ae103390bf04d2a00a41bf08d376dcdc497338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T15:47:09Z\\\",\\\"message\\\":\\\"2025-12-06T15:46:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5f77eacc-42b0-4fed-a916-5e52a3f47837\\\\n2025-12-06T15:46:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5f77eacc-42b0-4fed-a916-5e52a3f47837 to /host/opt/cni/bin/\\\\n2025-12-06T15:46:24Z [verbose] multus-daemon started\\\\n2025-12-06T15:46:24Z [verbose] Readiness Indicator file check\\\\n2025-12-06T15:47:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:47:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5246p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gl495\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.634586 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.634629 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.634645 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.634667 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.634684 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:40Z","lastTransitionTime":"2025-12-06T15:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.637038 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57b28270-8798-4400-8a2d-822cb111282d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://003193c76d703b082f22efd1dd14bcb6e02c9bc802af8383e98d826e933a9fc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9223ef169f965d4e9f428b51a46333754cf93cb5e138d8178978ccffb7a7cf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e448d353d75018454ed4933a9c59733209cd57a0417678dffe2f4b1b84899bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81d9773f0d32a4139def2222f13e67ff925a8428f37744ac0884c5fd8e96206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04ad638ccb246b435e3f06aafe6c60a26f59f7a4f46ceedb593a0ef2c1d43aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd6bd0864651c94d066ca08a4b86cbaeca491673efbb44153d871ce680ff7207\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b222d022d50bc54ed1752b54ccc849946790bdf443674e2a7b8f395d970fd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b9067dee557c9b4f75d6e6086a429d87acd282d8ef173db0d91c34cc0e1be76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.655135 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9de44d2-42dc-46fb-9780-4dcd9e5f9c86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ab1d9f15bba233c4bbd04ece144a5a8119c168abaeee7d4544269b55ad58ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b14a8cf06c49b6922327f25905466f86e1a0fd0a0b852f751bac6f8cd7db9e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvwp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bd4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.672910 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kzbhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18a94103-5ed7-4e27-a58d-7c989b0f70f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d8s4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:35Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kzbhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.692575 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da167c75-90a5-469c-b6f7-135c2713f69a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b36a4cf41bf4b2c7cea2c3931d55765fbfecc46a372dd995bfc46190647119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3a77641015b85b8ae2f0ac88680eccee3a480b918610f6329a615756a8f64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e3631fd444be9464339c6467a2ebfa856d503a00347de7efab6ec26b1198d7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.714519 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ed565fa213413fb136776bad1903e977bbc678e964383299787b93c11e23b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.737911 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.737973 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.737990 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.738016 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.738034 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:40Z","lastTransitionTime":"2025-12-06T15:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.760198 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.781311 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.798863 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mhncd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dd11f51-430a-4e7a-ab55-9666abfb1e82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7442065164fb3f4fc4f5d12645104a42c14fd7a4fd04aeb2fb46188e90983809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6ddx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mhncd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.822051 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e54e31f-e9f8-4dd5-8738-c785df002c93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:21Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b654f2820e7e70a18617e577d1dffa10f2e6e786d8e8a2483493d1e2cbf32e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b654f2820e7e70a18617e577d1dffa10f2e6e786d8e8a2483493d1e2cbf32e6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T15:47:18Z\\\",\\\"message\\\":\\\"4-c6801b70dcb0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 15:47:18.251613 6661 services_controller.go:443] Built service openshift-image-registry/image-registry LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.93\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:5000, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1206 15:47:18.251617 6661 ovn.go:134] Ensuring zone local for Pod openshift-kube-scheduler/openshift-kube-scheduler-crc in node crc\\\\nI1206 15:47:18.251627 6661 services_controller.go:444] Built service openshift-image-registry/image-registry LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1206 15:47:18.251637 6661 services_controller.go:445] Built service openshift-image-registry/image-registry LB template configs for network=default: []services.lbConfig(nil)\\\\nF1206 15:47:18.249639 6661 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T15:47:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xdt4d_openshift-ovn-kubernetes(0e54e31f-e9f8-4dd5-8738-c785df002c93)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xdt4d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.833645 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-frdb4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d9dc77-2a52-4c06-a7c9-ac70fdc1e2c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96883dc056994b3d5affe2741751a9b12754d19ce228b8916d9d87beab654bc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgn4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-frdb4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.840105 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.840130 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.840140 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.840176 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.840186 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:40Z","lastTransitionTime":"2025-12-06T15:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.850137 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"393f3b4b-ef95-4f70-af75-9fa123e0d52b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3d7272d34a92fa558ed95d304ebaf3f6f0e79e5d60538187542d201250d93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e297ada3f58d22f81f4014c0d47f04b0fe2e68c1e032e9b17e85907f27f9a6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e297ada3f58d22f81f4014c0d47f04b0fe2e68c1e032e9b17e85907f27f9a6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.868630 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"012e3f9a-67f1-45f2-8e57-76928bb4d5c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a77b5cb2e8c50b39c3369f156a7800603e7d8bf3d6a544ad0f3a6814361338f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2f1dd12351d9f1ce3ccc9711361cf346d9aa1c4cb5d971020e242c80e557eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8326b53a1a99f7cbd09d55008e9c1b4087993a25ae4b50fafc70e4c7b99138d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9471d3db940ee579d5b60cf1f7750c2c265a8f8e549aaa7ed60e82982cf80d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9471d3db940ee579d5b60cf1f7750c2c265a8f8e549aaa7ed60e82982cf80d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T15:46:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T15:46:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.888908 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.907082 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d88e8bae-c055-4c55-b548-f621ff96de06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T15:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://281cb78215285fddfa12523e8a13c2adc9adb51b35d64ec66e5259369b2af96c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T15:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cg64b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T15:46:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t5xp8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T15:47:40Z is after 2025-08-24T17:21:41Z" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.942655 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.942700 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.942720 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.942747 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:40 crc kubenswrapper[4813]: I1206 15:47:40.942771 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:40Z","lastTransitionTime":"2025-12-06T15:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.045513 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.045560 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.045571 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.045586 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.045598 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:41Z","lastTransitionTime":"2025-12-06T15:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.148889 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.148962 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.148980 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.149053 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.149082 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:41Z","lastTransitionTime":"2025-12-06T15:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.251776 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.251843 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.251861 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.251889 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.251908 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:41Z","lastTransitionTime":"2025-12-06T15:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.354890 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.354930 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.354939 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.354953 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.354966 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:41Z","lastTransitionTime":"2025-12-06T15:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.457752 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.457794 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.457805 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.457823 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.457834 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:41Z","lastTransitionTime":"2025-12-06T15:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.486711 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:41 crc kubenswrapper[4813]: E1206 15:47:41.486868 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.487113 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:41 crc kubenswrapper[4813]: E1206 15:47:41.487219 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.487435 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:41 crc kubenswrapper[4813]: E1206 15:47:41.487528 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.487709 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:41 crc kubenswrapper[4813]: E1206 15:47:41.487800 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.560854 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.560901 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.560918 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.560943 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.560962 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:41Z","lastTransitionTime":"2025-12-06T15:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.663636 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.663682 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.663699 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.663720 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.663737 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:41Z","lastTransitionTime":"2025-12-06T15:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.783726 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.783770 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.783782 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.783801 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.783814 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:41Z","lastTransitionTime":"2025-12-06T15:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.886720 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.886793 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.886812 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.886837 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.886873 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:41Z","lastTransitionTime":"2025-12-06T15:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.989931 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.989980 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.989998 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.990020 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:41 crc kubenswrapper[4813]: I1206 15:47:41.990037 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:41Z","lastTransitionTime":"2025-12-06T15:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.093330 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.093386 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.093404 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.093427 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.093443 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:42Z","lastTransitionTime":"2025-12-06T15:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.196426 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.196484 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.196502 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.196527 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.196547 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:42Z","lastTransitionTime":"2025-12-06T15:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.298869 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.299247 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.299418 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.299591 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.299618 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:42Z","lastTransitionTime":"2025-12-06T15:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.401945 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.402002 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.402020 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.402043 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.402060 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:42Z","lastTransitionTime":"2025-12-06T15:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.504919 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.504952 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.504961 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.504974 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.504983 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:42Z","lastTransitionTime":"2025-12-06T15:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.607324 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.608053 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.608196 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.608383 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.608536 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:42Z","lastTransitionTime":"2025-12-06T15:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.711526 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.711599 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.711618 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.711647 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.711670 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:42Z","lastTransitionTime":"2025-12-06T15:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.814817 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.814866 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.814882 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.814903 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.814921 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:42Z","lastTransitionTime":"2025-12-06T15:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.917037 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.917083 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.917100 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.917120 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:42 crc kubenswrapper[4813]: I1206 15:47:42.917137 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:42Z","lastTransitionTime":"2025-12-06T15:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.019687 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.019725 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.019739 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.019760 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.019775 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:43Z","lastTransitionTime":"2025-12-06T15:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.121978 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.122312 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.122398 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.122495 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.122626 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:43Z","lastTransitionTime":"2025-12-06T15:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.225380 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.225445 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.225455 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.225472 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.225483 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:43Z","lastTransitionTime":"2025-12-06T15:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.328058 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.328115 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.328125 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.328137 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.328146 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:43Z","lastTransitionTime":"2025-12-06T15:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.430446 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.430511 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.430528 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.430555 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.430573 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:43Z","lastTransitionTime":"2025-12-06T15:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.485919 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.486081 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:43 crc kubenswrapper[4813]: E1206 15:47:43.486213 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.486321 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:43 crc kubenswrapper[4813]: E1206 15:47:43.486427 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:47:43 crc kubenswrapper[4813]: E1206 15:47:43.486547 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.486750 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:43 crc kubenswrapper[4813]: E1206 15:47:43.487026 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.533851 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.533901 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.533915 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.533937 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.533972 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:43Z","lastTransitionTime":"2025-12-06T15:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.636322 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.636397 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.636419 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.636440 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.636457 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:43Z","lastTransitionTime":"2025-12-06T15:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.739989 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.740044 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.740062 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.740086 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.740103 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:43Z","lastTransitionTime":"2025-12-06T15:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.843157 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.843524 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.843664 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.843807 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.843931 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:43Z","lastTransitionTime":"2025-12-06T15:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.947717 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.947769 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.947788 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.947812 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:43 crc kubenswrapper[4813]: I1206 15:47:43.947830 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:43Z","lastTransitionTime":"2025-12-06T15:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.051109 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.051154 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.051172 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.051190 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.051203 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:44Z","lastTransitionTime":"2025-12-06T15:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.153990 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.154052 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.154070 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.154098 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.154116 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:44Z","lastTransitionTime":"2025-12-06T15:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.256392 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.256469 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.256494 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.256523 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.256544 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:44Z","lastTransitionTime":"2025-12-06T15:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.360470 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.361455 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.361524 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.361552 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.361585 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:44Z","lastTransitionTime":"2025-12-06T15:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.464423 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.464511 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.464529 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.464553 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.464571 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:44Z","lastTransitionTime":"2025-12-06T15:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.486702 4813 scope.go:117] "RemoveContainer" containerID="6b654f2820e7e70a18617e577d1dffa10f2e6e786d8e8a2483493d1e2cbf32e6" Dec 06 15:47:44 crc kubenswrapper[4813]: E1206 15:47:44.486859 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xdt4d_openshift-ovn-kubernetes(0e54e31f-e9f8-4dd5-8738-c785df002c93)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.568897 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.569022 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.569044 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.569113 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.569136 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:44Z","lastTransitionTime":"2025-12-06T15:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.671877 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.671932 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.671944 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.671962 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.672363 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:44Z","lastTransitionTime":"2025-12-06T15:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.775551 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.775870 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.776070 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.776215 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.776563 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:44Z","lastTransitionTime":"2025-12-06T15:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.879586 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.879650 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.879675 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.879705 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.879727 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:44Z","lastTransitionTime":"2025-12-06T15:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.982737 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.983140 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.983528 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.983900 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:44 crc kubenswrapper[4813]: I1206 15:47:44.984041 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:44Z","lastTransitionTime":"2025-12-06T15:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.086943 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.087309 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.087679 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.088079 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.088413 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:45Z","lastTransitionTime":"2025-12-06T15:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.190920 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.191517 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.191674 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.191809 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.191925 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:45Z","lastTransitionTime":"2025-12-06T15:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.294836 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.295084 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.295244 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.295392 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.295479 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:45Z","lastTransitionTime":"2025-12-06T15:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.398358 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.398420 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.398440 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.398465 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.398484 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:45Z","lastTransitionTime":"2025-12-06T15:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.486424 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.486445 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:45 crc kubenswrapper[4813]: E1206 15:47:45.486571 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.486733 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.486875 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:45 crc kubenswrapper[4813]: E1206 15:47:45.486868 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:47:45 crc kubenswrapper[4813]: E1206 15:47:45.486995 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:47:45 crc kubenswrapper[4813]: E1206 15:47:45.487071 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.509961 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.510499 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.510683 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.511319 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.511493 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:45Z","lastTransitionTime":"2025-12-06T15:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.616173 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.616578 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.616827 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.617039 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.617205 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:45Z","lastTransitionTime":"2025-12-06T15:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.720206 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.720295 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.720319 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.720349 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.720370 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:45Z","lastTransitionTime":"2025-12-06T15:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.823093 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.823150 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.823167 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.823191 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.823209 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:45Z","lastTransitionTime":"2025-12-06T15:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.926540 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.926591 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.926607 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.926634 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:45 crc kubenswrapper[4813]: I1206 15:47:45.926651 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:45Z","lastTransitionTime":"2025-12-06T15:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.029022 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.029328 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.029410 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.029489 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.029557 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:46Z","lastTransitionTime":"2025-12-06T15:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.132332 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.132378 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.132395 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.132418 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.132435 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:46Z","lastTransitionTime":"2025-12-06T15:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.235594 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.235797 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.235828 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.235857 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.235880 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:46Z","lastTransitionTime":"2025-12-06T15:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.338544 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.338629 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.338683 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.338710 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.338726 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:46Z","lastTransitionTime":"2025-12-06T15:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.441510 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.441579 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.441628 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.441659 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.441677 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:46Z","lastTransitionTime":"2025-12-06T15:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.544167 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.544234 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.544252 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.544314 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.544333 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:46Z","lastTransitionTime":"2025-12-06T15:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.647211 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.647295 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.647313 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.647336 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.647352 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:46Z","lastTransitionTime":"2025-12-06T15:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.751459 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.751490 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.751499 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.751511 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.751519 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:46Z","lastTransitionTime":"2025-12-06T15:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.854085 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.854166 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.854193 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.854229 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.854254 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:46Z","lastTransitionTime":"2025-12-06T15:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.957406 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.957687 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.957835 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.957974 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:46 crc kubenswrapper[4813]: I1206 15:47:46.958100 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:46Z","lastTransitionTime":"2025-12-06T15:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.061378 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.061445 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.061468 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.061497 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.061521 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:47Z","lastTransitionTime":"2025-12-06T15:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.164863 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.165029 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.165050 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.165291 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.165311 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:47Z","lastTransitionTime":"2025-12-06T15:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.268743 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.268787 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.268804 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.268826 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.268842 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:47Z","lastTransitionTime":"2025-12-06T15:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.372372 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.372426 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.372481 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.372509 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.372532 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:47Z","lastTransitionTime":"2025-12-06T15:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.475355 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.475407 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.475429 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.475452 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.475472 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:47Z","lastTransitionTime":"2025-12-06T15:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.486424 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:47 crc kubenswrapper[4813]: E1206 15:47:47.486555 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.486761 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:47 crc kubenswrapper[4813]: E1206 15:47:47.486859 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.487063 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:47 crc kubenswrapper[4813]: E1206 15:47:47.487164 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.487418 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:47 crc kubenswrapper[4813]: E1206 15:47:47.487544 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.578572 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.578624 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.578641 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.578663 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.578680 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:47Z","lastTransitionTime":"2025-12-06T15:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.623146 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.623221 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.623247 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.623318 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.623343 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T15:47:47Z","lastTransitionTime":"2025-12-06T15:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.703923 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5k6s"] Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.704768 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5k6s" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.707900 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.708025 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.708183 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.708345 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.767057 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-mhncd" podStartSLOduration=88.767034722 podStartE2EDuration="1m28.767034722s" podCreationTimestamp="2025-12-06 15:46:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:47:47.740757216 +0000 UTC m=+107.631636792" watchObservedRunningTime="2025-12-06 15:47:47.767034722 +0000 UTC m=+107.657914308" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.779039 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-frdb4" podStartSLOduration=88.779017827 podStartE2EDuration="1m28.779017827s" podCreationTimestamp="2025-12-06 15:46:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:47:47.778729739 +0000 UTC m=+107.669609325" watchObservedRunningTime="2025-12-06 15:47:47.779017827 +0000 UTC m=+107.669897403" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.815184 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a9c2f3a9-1291-461f-a477-71ae4e48eee5-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-g5k6s\" (UID: \"a9c2f3a9-1291-461f-a477-71ae4e48eee5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5k6s" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.815189 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=88.815169329 podStartE2EDuration="1m28.815169329s" podCreationTimestamp="2025-12-06 15:46:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:47:47.814358146 +0000 UTC m=+107.705237722" watchObservedRunningTime="2025-12-06 15:47:47.815169329 +0000 UTC m=+107.706048945" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.815250 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/a9c2f3a9-1291-461f-a477-71ae4e48eee5-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-g5k6s\" (UID: \"a9c2f3a9-1291-461f-a477-71ae4e48eee5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5k6s" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.815343 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/a9c2f3a9-1291-461f-a477-71ae4e48eee5-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-g5k6s\" (UID: \"a9c2f3a9-1291-461f-a477-71ae4e48eee5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5k6s" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.815403 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a9c2f3a9-1291-461f-a477-71ae4e48eee5-service-ca\") pod \"cluster-version-operator-5c965bbfc6-g5k6s\" (UID: \"a9c2f3a9-1291-461f-a477-71ae4e48eee5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5k6s" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.815434 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9c2f3a9-1291-461f-a477-71ae4e48eee5-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-g5k6s\" (UID: \"a9c2f3a9-1291-461f-a477-71ae4e48eee5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5k6s" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.875913 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podStartSLOduration=88.875886708 podStartE2EDuration="1m28.875886708s" podCreationTimestamp="2025-12-06 15:46:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:47:47.861913737 +0000 UTC m=+107.752793333" watchObservedRunningTime="2025-12-06 15:47:47.875886708 +0000 UTC m=+107.766766314" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.876122 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=35.876113894 podStartE2EDuration="35.876113894s" podCreationTimestamp="2025-12-06 15:47:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:47:47.874998813 +0000 UTC m=+107.765878399" watchObservedRunningTime="2025-12-06 15:47:47.876113894 +0000 UTC m=+107.766993500" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.890480 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=57.890460286 podStartE2EDuration="57.890460286s" podCreationTimestamp="2025-12-06 15:46:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:47:47.890092225 +0000 UTC m=+107.780971841" watchObservedRunningTime="2025-12-06 15:47:47.890460286 +0000 UTC m=+107.781339892" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.916016 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a9c2f3a9-1291-461f-a477-71ae4e48eee5-service-ca\") pod \"cluster-version-operator-5c965bbfc6-g5k6s\" (UID: \"a9c2f3a9-1291-461f-a477-71ae4e48eee5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5k6s" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.916074 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9c2f3a9-1291-461f-a477-71ae4e48eee5-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-g5k6s\" (UID: \"a9c2f3a9-1291-461f-a477-71ae4e48eee5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5k6s" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.916127 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a9c2f3a9-1291-461f-a477-71ae4e48eee5-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-g5k6s\" (UID: \"a9c2f3a9-1291-461f-a477-71ae4e48eee5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5k6s" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.916150 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/a9c2f3a9-1291-461f-a477-71ae4e48eee5-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-g5k6s\" (UID: \"a9c2f3a9-1291-461f-a477-71ae4e48eee5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5k6s" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.916194 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/a9c2f3a9-1291-461f-a477-71ae4e48eee5-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-g5k6s\" (UID: \"a9c2f3a9-1291-461f-a477-71ae4e48eee5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5k6s" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.916250 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/a9c2f3a9-1291-461f-a477-71ae4e48eee5-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-g5k6s\" (UID: \"a9c2f3a9-1291-461f-a477-71ae4e48eee5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5k6s" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.917138 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a9c2f3a9-1291-461f-a477-71ae4e48eee5-service-ca\") pod \"cluster-version-operator-5c965bbfc6-g5k6s\" (UID: \"a9c2f3a9-1291-461f-a477-71ae4e48eee5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5k6s" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.917866 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/a9c2f3a9-1291-461f-a477-71ae4e48eee5-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-g5k6s\" (UID: \"a9c2f3a9-1291-461f-a477-71ae4e48eee5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5k6s" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.929653 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-t9zmh" podStartSLOduration=88.929632872 podStartE2EDuration="1m28.929632872s" podCreationTimestamp="2025-12-06 15:46:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:47:47.929477997 +0000 UTC m=+107.820357613" watchObservedRunningTime="2025-12-06 15:47:47.929632872 +0000 UTC m=+107.820512458" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.936727 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9c2f3a9-1291-461f-a477-71ae4e48eee5-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-g5k6s\" (UID: \"a9c2f3a9-1291-461f-a477-71ae4e48eee5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5k6s" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.945064 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-gl495" podStartSLOduration=88.945045463 podStartE2EDuration="1m28.945045463s" podCreationTimestamp="2025-12-06 15:46:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:47:47.942214064 +0000 UTC m=+107.833093660" watchObservedRunningTime="2025-12-06 15:47:47.945045463 +0000 UTC m=+107.835925049" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.945872 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a9c2f3a9-1291-461f-a477-71ae4e48eee5-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-g5k6s\" (UID: \"a9c2f3a9-1291-461f-a477-71ae4e48eee5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5k6s" Dec 06 15:47:47 crc kubenswrapper[4813]: I1206 15:47:47.964704 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=86.964684673 podStartE2EDuration="1m26.964684673s" podCreationTimestamp="2025-12-06 15:46:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:47:47.963248883 +0000 UTC m=+107.854128469" watchObservedRunningTime="2025-12-06 15:47:47.964684673 +0000 UTC m=+107.855564259" Dec 06 15:47:48 crc kubenswrapper[4813]: I1206 15:47:48.027865 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5k6s" Dec 06 15:47:48 crc kubenswrapper[4813]: I1206 15:47:48.033410 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=86.033397506 podStartE2EDuration="1m26.033397506s" podCreationTimestamp="2025-12-06 15:46:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:47:48.032889261 +0000 UTC m=+107.923768857" watchObservedRunningTime="2025-12-06 15:47:48.033397506 +0000 UTC m=+107.924277082" Dec 06 15:47:48 crc kubenswrapper[4813]: I1206 15:47:48.059126 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bd4fr" podStartSLOduration=88.059107585 podStartE2EDuration="1m28.059107585s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:47:48.058453777 +0000 UTC m=+107.949333373" watchObservedRunningTime="2025-12-06 15:47:48.059107585 +0000 UTC m=+107.949987171" Dec 06 15:47:48 crc kubenswrapper[4813]: I1206 15:47:48.706132 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5k6s" event={"ID":"a9c2f3a9-1291-461f-a477-71ae4e48eee5","Type":"ContainerStarted","Data":"8f0fc7fe5de0314a0e8250bede83fbc4f2a2c94fa693cef0f337b5ed9c67126d"} Dec 06 15:47:48 crc kubenswrapper[4813]: I1206 15:47:48.706207 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5k6s" event={"ID":"a9c2f3a9-1291-461f-a477-71ae4e48eee5","Type":"ContainerStarted","Data":"f8d91cf1e98ea382c13c7749ba6fb22512c84337fa820f036fa6ce6d7a05bf67"} Dec 06 15:47:48 crc kubenswrapper[4813]: I1206 15:47:48.726173 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5k6s" podStartSLOduration=89.726144591 podStartE2EDuration="1m29.726144591s" podCreationTimestamp="2025-12-06 15:46:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:47:48.725623777 +0000 UTC m=+108.616503353" watchObservedRunningTime="2025-12-06 15:47:48.726144591 +0000 UTC m=+108.617024207" Dec 06 15:47:49 crc kubenswrapper[4813]: I1206 15:47:49.486711 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:49 crc kubenswrapper[4813]: E1206 15:47:49.486876 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:47:49 crc kubenswrapper[4813]: I1206 15:47:49.487011 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:49 crc kubenswrapper[4813]: I1206 15:47:49.487137 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:49 crc kubenswrapper[4813]: I1206 15:47:49.487420 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:49 crc kubenswrapper[4813]: E1206 15:47:49.487377 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:47:49 crc kubenswrapper[4813]: E1206 15:47:49.487466 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:47:49 crc kubenswrapper[4813]: E1206 15:47:49.487506 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:47:51 crc kubenswrapper[4813]: I1206 15:47:51.486325 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:51 crc kubenswrapper[4813]: I1206 15:47:51.486404 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:51 crc kubenswrapper[4813]: I1206 15:47:51.486404 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:51 crc kubenswrapper[4813]: I1206 15:47:51.486468 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:51 crc kubenswrapper[4813]: E1206 15:47:51.486808 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:47:51 crc kubenswrapper[4813]: E1206 15:47:51.487321 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:47:51 crc kubenswrapper[4813]: E1206 15:47:51.487457 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:47:51 crc kubenswrapper[4813]: E1206 15:47:51.487175 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:47:53 crc kubenswrapper[4813]: I1206 15:47:53.485776 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:53 crc kubenswrapper[4813]: I1206 15:47:53.485833 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:53 crc kubenswrapper[4813]: I1206 15:47:53.485826 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:53 crc kubenswrapper[4813]: I1206 15:47:53.485796 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:53 crc kubenswrapper[4813]: E1206 15:47:53.485935 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:47:53 crc kubenswrapper[4813]: E1206 15:47:53.486142 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:47:53 crc kubenswrapper[4813]: E1206 15:47:53.486184 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:47:53 crc kubenswrapper[4813]: E1206 15:47:53.486248 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:47:55 crc kubenswrapper[4813]: I1206 15:47:55.486210 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:55 crc kubenswrapper[4813]: I1206 15:47:55.486342 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:55 crc kubenswrapper[4813]: E1206 15:47:55.486546 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:47:55 crc kubenswrapper[4813]: I1206 15:47:55.486594 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:55 crc kubenswrapper[4813]: E1206 15:47:55.486819 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:47:55 crc kubenswrapper[4813]: I1206 15:47:55.486241 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:55 crc kubenswrapper[4813]: E1206 15:47:55.487782 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:47:55 crc kubenswrapper[4813]: E1206 15:47:55.488319 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:47:56 crc kubenswrapper[4813]: I1206 15:47:56.487721 4813 scope.go:117] "RemoveContainer" containerID="6b654f2820e7e70a18617e577d1dffa10f2e6e786d8e8a2483493d1e2cbf32e6" Dec 06 15:47:56 crc kubenswrapper[4813]: E1206 15:47:56.487989 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xdt4d_openshift-ovn-kubernetes(0e54e31f-e9f8-4dd5-8738-c785df002c93)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" Dec 06 15:47:56 crc kubenswrapper[4813]: I1206 15:47:56.739837 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gl495_856bb45b-525c-4ef0-bf7c-0691cf54b342/kube-multus/1.log" Dec 06 15:47:56 crc kubenswrapper[4813]: I1206 15:47:56.740574 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gl495_856bb45b-525c-4ef0-bf7c-0691cf54b342/kube-multus/0.log" Dec 06 15:47:56 crc kubenswrapper[4813]: I1206 15:47:56.740607 4813 generic.go:334] "Generic (PLEG): container finished" podID="856bb45b-525c-4ef0-bf7c-0691cf54b342" containerID="7001bdf05c6acf52c9be35bd63ae103390bf04d2a00a41bf08d376dcdc497338" exitCode=1 Dec 06 15:47:56 crc kubenswrapper[4813]: I1206 15:47:56.740637 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gl495" event={"ID":"856bb45b-525c-4ef0-bf7c-0691cf54b342","Type":"ContainerDied","Data":"7001bdf05c6acf52c9be35bd63ae103390bf04d2a00a41bf08d376dcdc497338"} Dec 06 15:47:56 crc kubenswrapper[4813]: I1206 15:47:56.740669 4813 scope.go:117] "RemoveContainer" containerID="e52cfb160ef030b63e33c1db58b2fbdc64df5fe7ccb66f42b4de558f13ba122c" Dec 06 15:47:56 crc kubenswrapper[4813]: I1206 15:47:56.741022 4813 scope.go:117] "RemoveContainer" containerID="7001bdf05c6acf52c9be35bd63ae103390bf04d2a00a41bf08d376dcdc497338" Dec 06 15:47:56 crc kubenswrapper[4813]: E1206 15:47:56.741149 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-gl495_openshift-multus(856bb45b-525c-4ef0-bf7c-0691cf54b342)\"" pod="openshift-multus/multus-gl495" podUID="856bb45b-525c-4ef0-bf7c-0691cf54b342" Dec 06 15:47:57 crc kubenswrapper[4813]: I1206 15:47:57.486030 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:57 crc kubenswrapper[4813]: I1206 15:47:57.486074 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:57 crc kubenswrapper[4813]: E1206 15:47:57.486135 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:47:57 crc kubenswrapper[4813]: I1206 15:47:57.486042 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:57 crc kubenswrapper[4813]: I1206 15:47:57.486288 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:57 crc kubenswrapper[4813]: E1206 15:47:57.486471 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:47:57 crc kubenswrapper[4813]: E1206 15:47:57.486555 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:47:57 crc kubenswrapper[4813]: E1206 15:47:57.486657 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:47:57 crc kubenswrapper[4813]: I1206 15:47:57.745694 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gl495_856bb45b-525c-4ef0-bf7c-0691cf54b342/kube-multus/1.log" Dec 06 15:47:59 crc kubenswrapper[4813]: I1206 15:47:59.486535 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:47:59 crc kubenswrapper[4813]: I1206 15:47:59.486580 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:47:59 crc kubenswrapper[4813]: I1206 15:47:59.486566 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:47:59 crc kubenswrapper[4813]: I1206 15:47:59.486552 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:47:59 crc kubenswrapper[4813]: E1206 15:47:59.486845 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:47:59 crc kubenswrapper[4813]: E1206 15:47:59.486724 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:47:59 crc kubenswrapper[4813]: E1206 15:47:59.486985 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:47:59 crc kubenswrapper[4813]: E1206 15:47:59.487077 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:48:00 crc kubenswrapper[4813]: E1206 15:48:00.466362 4813 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 06 15:48:00 crc kubenswrapper[4813]: E1206 15:48:00.718885 4813 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 15:48:01 crc kubenswrapper[4813]: I1206 15:48:01.486644 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:48:01 crc kubenswrapper[4813]: I1206 15:48:01.486669 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:48:01 crc kubenswrapper[4813]: I1206 15:48:01.486738 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:48:01 crc kubenswrapper[4813]: I1206 15:48:01.486761 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:48:01 crc kubenswrapper[4813]: E1206 15:48:01.486913 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:48:01 crc kubenswrapper[4813]: E1206 15:48:01.487084 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:48:01 crc kubenswrapper[4813]: E1206 15:48:01.487309 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:48:01 crc kubenswrapper[4813]: E1206 15:48:01.487404 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:48:03 crc kubenswrapper[4813]: I1206 15:48:03.486746 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:48:03 crc kubenswrapper[4813]: I1206 15:48:03.486837 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:48:03 crc kubenswrapper[4813]: I1206 15:48:03.486788 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:48:03 crc kubenswrapper[4813]: I1206 15:48:03.486754 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:48:03 crc kubenswrapper[4813]: E1206 15:48:03.486974 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:48:03 crc kubenswrapper[4813]: E1206 15:48:03.487091 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:48:03 crc kubenswrapper[4813]: E1206 15:48:03.487238 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:48:03 crc kubenswrapper[4813]: E1206 15:48:03.487399 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:48:05 crc kubenswrapper[4813]: I1206 15:48:05.486218 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:48:05 crc kubenswrapper[4813]: I1206 15:48:05.486256 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:48:05 crc kubenswrapper[4813]: E1206 15:48:05.486438 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:48:05 crc kubenswrapper[4813]: E1206 15:48:05.486568 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:48:05 crc kubenswrapper[4813]: I1206 15:48:05.487502 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:48:05 crc kubenswrapper[4813]: I1206 15:48:05.487639 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:48:05 crc kubenswrapper[4813]: E1206 15:48:05.487762 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:48:05 crc kubenswrapper[4813]: E1206 15:48:05.488028 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:48:05 crc kubenswrapper[4813]: E1206 15:48:05.720781 4813 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 15:48:07 crc kubenswrapper[4813]: I1206 15:48:07.486772 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:48:07 crc kubenswrapper[4813]: E1206 15:48:07.486966 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:48:07 crc kubenswrapper[4813]: I1206 15:48:07.487035 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:48:07 crc kubenswrapper[4813]: I1206 15:48:07.487046 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:48:07 crc kubenswrapper[4813]: E1206 15:48:07.487129 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:48:07 crc kubenswrapper[4813]: E1206 15:48:07.487412 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:48:07 crc kubenswrapper[4813]: I1206 15:48:07.486957 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:48:07 crc kubenswrapper[4813]: E1206 15:48:07.488097 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:48:07 crc kubenswrapper[4813]: I1206 15:48:07.488801 4813 scope.go:117] "RemoveContainer" containerID="6b654f2820e7e70a18617e577d1dffa10f2e6e786d8e8a2483493d1e2cbf32e6" Dec 06 15:48:07 crc kubenswrapper[4813]: I1206 15:48:07.815514 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xdt4d_0e54e31f-e9f8-4dd5-8738-c785df002c93/ovnkube-controller/3.log" Dec 06 15:48:07 crc kubenswrapper[4813]: I1206 15:48:07.817760 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" event={"ID":"0e54e31f-e9f8-4dd5-8738-c785df002c93","Type":"ContainerStarted","Data":"5cf32df4a99c42b7b090412102796d4a6d265bc5a4de2ecf1648cadbac28ea73"} Dec 06 15:48:07 crc kubenswrapper[4813]: I1206 15:48:07.818816 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:48:08 crc kubenswrapper[4813]: I1206 15:48:08.277534 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" podStartSLOduration=108.27751336 podStartE2EDuration="1m48.27751336s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:07.845952693 +0000 UTC m=+127.736832309" watchObservedRunningTime="2025-12-06 15:48:08.27751336 +0000 UTC m=+128.168392936" Dec 06 15:48:08 crc kubenswrapper[4813]: I1206 15:48:08.278476 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-kzbhb"] Dec 06 15:48:08 crc kubenswrapper[4813]: I1206 15:48:08.278593 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:48:08 crc kubenswrapper[4813]: E1206 15:48:08.278701 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:48:09 crc kubenswrapper[4813]: I1206 15:48:09.486102 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:48:09 crc kubenswrapper[4813]: I1206 15:48:09.486175 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:48:09 crc kubenswrapper[4813]: E1206 15:48:09.486347 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:48:09 crc kubenswrapper[4813]: E1206 15:48:09.486474 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:48:09 crc kubenswrapper[4813]: I1206 15:48:09.486537 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:48:09 crc kubenswrapper[4813]: E1206 15:48:09.486803 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:48:10 crc kubenswrapper[4813]: I1206 15:48:10.488557 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:48:10 crc kubenswrapper[4813]: E1206 15:48:10.489411 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:48:10 crc kubenswrapper[4813]: I1206 15:48:10.489676 4813 scope.go:117] "RemoveContainer" containerID="7001bdf05c6acf52c9be35bd63ae103390bf04d2a00a41bf08d376dcdc497338" Dec 06 15:48:10 crc kubenswrapper[4813]: E1206 15:48:10.723501 4813 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 15:48:11 crc kubenswrapper[4813]: I1206 15:48:11.486399 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:48:11 crc kubenswrapper[4813]: E1206 15:48:11.486570 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:48:11 crc kubenswrapper[4813]: I1206 15:48:11.486663 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:48:11 crc kubenswrapper[4813]: E1206 15:48:11.486747 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:48:11 crc kubenswrapper[4813]: I1206 15:48:11.486808 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:48:11 crc kubenswrapper[4813]: E1206 15:48:11.486895 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:48:11 crc kubenswrapper[4813]: I1206 15:48:11.834000 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gl495_856bb45b-525c-4ef0-bf7c-0691cf54b342/kube-multus/1.log" Dec 06 15:48:11 crc kubenswrapper[4813]: I1206 15:48:11.834377 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gl495" event={"ID":"856bb45b-525c-4ef0-bf7c-0691cf54b342","Type":"ContainerStarted","Data":"ef40f8e7e7629e97c914ef68c7d15da2e144cc9558102b30198efb0618912fe8"} Dec 06 15:48:12 crc kubenswrapper[4813]: I1206 15:48:12.486448 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:48:12 crc kubenswrapper[4813]: E1206 15:48:12.486675 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:48:13 crc kubenswrapper[4813]: I1206 15:48:13.486199 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:48:13 crc kubenswrapper[4813]: I1206 15:48:13.486201 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:48:13 crc kubenswrapper[4813]: E1206 15:48:13.487291 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:48:13 crc kubenswrapper[4813]: I1206 15:48:13.486229 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:48:13 crc kubenswrapper[4813]: E1206 15:48:13.487424 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:48:13 crc kubenswrapper[4813]: E1206 15:48:13.487602 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:48:14 crc kubenswrapper[4813]: I1206 15:48:14.486349 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:48:14 crc kubenswrapper[4813]: E1206 15:48:14.486544 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kzbhb" podUID="18a94103-5ed7-4e27-a58d-7c989b0f70f3" Dec 06 15:48:15 crc kubenswrapper[4813]: I1206 15:48:15.486059 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:48:15 crc kubenswrapper[4813]: I1206 15:48:15.486099 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:48:15 crc kubenswrapper[4813]: I1206 15:48:15.486060 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:48:15 crc kubenswrapper[4813]: E1206 15:48:15.486243 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 15:48:15 crc kubenswrapper[4813]: E1206 15:48:15.486380 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 15:48:15 crc kubenswrapper[4813]: E1206 15:48:15.486559 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 15:48:16 crc kubenswrapper[4813]: I1206 15:48:16.486979 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:48:16 crc kubenswrapper[4813]: I1206 15:48:16.491212 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 06 15:48:16 crc kubenswrapper[4813]: I1206 15:48:16.491518 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 06 15:48:17 crc kubenswrapper[4813]: I1206 15:48:17.486493 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:48:17 crc kubenswrapper[4813]: I1206 15:48:17.486619 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:48:17 crc kubenswrapper[4813]: I1206 15:48:17.486670 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:48:17 crc kubenswrapper[4813]: I1206 15:48:17.488708 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 06 15:48:17 crc kubenswrapper[4813]: I1206 15:48:17.489455 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 06 15:48:17 crc kubenswrapper[4813]: I1206 15:48:17.489652 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 06 15:48:17 crc kubenswrapper[4813]: I1206 15:48:17.490044 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.407601 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.454436 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6bbs5"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.454938 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-6bbs5" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.457018 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.457562 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.460684 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-dq6hp"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.461433 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-dq6hp" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.462315 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fszk7"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.462744 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.462802 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.462929 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.463197 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.463917 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.464550 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.475181 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.477857 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.478171 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.478370 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.478432 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.478688 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.478879 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.479167 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.479322 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.479429 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.479721 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.479772 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.479854 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.479950 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.479986 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.480099 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.480823 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.481017 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.481292 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 06 15:48:18 crc kubenswrapper[4813]: W1206 15:48:18.481318 4813 reflector.go:561] object-"openshift-authentication"/"v4-0-config-system-service-ca": failed to list *v1.ConfigMap: configmaps "v4-0-config-system-service-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Dec 06 15:48:18 crc kubenswrapper[4813]: E1206 15:48:18.481366 4813 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-service-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"v4-0-config-system-service-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.481415 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.481619 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.481720 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.481804 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.481853 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.482028 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.482084 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-klqt2"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.483028 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-klqt2" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.484116 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-5t7m6"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.494576 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-ktnmq"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.499881 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.500760 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.501594 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.502000 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-5t7m6" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.502324 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.502510 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktnmq" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.507056 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.507207 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.507396 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.512514 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.512805 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.513762 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.515869 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.517451 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.517910 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8vzrm"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.518628 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.519434 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.519484 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.519550 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.528108 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rtsv4"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.528648 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-gb6cc"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.530719 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pmr8t"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.530990 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6bbs5"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.531084 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pmr8t" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.531614 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-dq6hp"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.531782 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rtsv4" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.532129 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-gb6cc" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.532791 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.542204 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.542520 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.542646 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.543533 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.543921 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.545436 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.545612 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.545712 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.546695 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.546732 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-klqt2"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.546743 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fszk7"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.546783 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.547019 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.547111 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.549208 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-kkbqh"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.549545 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d4h8p"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.549890 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d4h8p" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.550215 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-kkbqh" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.556568 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.558354 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.558520 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.558591 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.558662 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.558731 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.558800 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.558888 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.558967 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.559043 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.559114 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.561504 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.561833 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.562353 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.562457 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.562573 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.562625 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.562633 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.562723 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.562743 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.562749 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.562859 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.562915 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.562961 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.562983 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.563046 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.563059 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.563147 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.579981 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.583444 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.592234 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzb98"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.595300 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.597052 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.599024 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.599718 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.600112 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.600718 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.600743 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.601018 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-j79t8"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.601539 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzb98" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.607682 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.612285 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-2twln"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.612416 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.618563 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.618881 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.623499 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/30e8f6c3-6c7e-45dc-9a6f-e2de13d1322c-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-pmr8t\" (UID: \"30e8f6c3-6c7e-45dc-9a6f-e2de13d1322c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pmr8t" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.623536 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.623561 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d57f80ec-52ac-409d-9dc1-93a89d59f714-serving-cert\") pod \"authentication-operator-69f744f599-5t7m6\" (UID: \"d57f80ec-52ac-409d-9dc1-93a89d59f714\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5t7m6" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.623586 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5qmj\" (UniqueName: \"kubernetes.io/projected/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-kube-api-access-l5qmj\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.623603 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/994c44ca-f520-4452-b267-961bd400908e-audit-dir\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.623647 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4b409627-4716-4bdd-a660-4f88693150a2-client-ca\") pod \"route-controller-manager-6576b87f9c-6896q\" (UID: \"4b409627-4716-4bdd-a660-4f88693150a2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.623668 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-serving-cert\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.623686 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a1353568-e317-4a7d-8765-675821dac72b-encryption-config\") pod \"apiserver-7bbb656c7d-b955p\" (UID: \"a1353568-e317-4a7d-8765-675821dac72b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.623703 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-etcd-client\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.623734 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.623753 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b409627-4716-4bdd-a660-4f88693150a2-config\") pod \"route-controller-manager-6576b87f9c-6896q\" (UID: \"4b409627-4716-4bdd-a660-4f88693150a2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.623770 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-image-import-ca\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.623789 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.623806 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a1353568-e317-4a7d-8765-675821dac72b-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-b955p\" (UID: \"a1353568-e317-4a7d-8765-675821dac72b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.623826 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2f02af08-d7ad-4460-b103-cf44d7598449-trusted-ca\") pod \"console-operator-58897d9998-klqt2\" (UID: \"2f02af08-d7ad-4460-b103-cf44d7598449\") " pod="openshift-console-operator/console-operator-58897d9998-klqt2" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.623864 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a1353568-e317-4a7d-8765-675821dac72b-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-b955p\" (UID: \"a1353568-e317-4a7d-8765-675821dac72b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.623886 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30e8f6c3-6c7e-45dc-9a6f-e2de13d1322c-config\") pod \"openshift-apiserver-operator-796bbdcf4f-pmr8t\" (UID: \"30e8f6c3-6c7e-45dc-9a6f-e2de13d1322c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pmr8t" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.623914 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.623932 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1d5678dc-7a74-4fff-818c-e34321037686-client-ca\") pod \"controller-manager-879f6c89f-6bbs5\" (UID: \"1d5678dc-7a74-4fff-818c-e34321037686\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6bbs5" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.623951 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.623970 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/36f25c5a-3e57-4367-b306-db9661e4f7c9-oauth-serving-cert\") pod \"console-f9d7485db-gb6cc\" (UID: \"36f25c5a-3e57-4367-b306-db9661e4f7c9\") " pod="openshift-console/console-f9d7485db-gb6cc" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.623988 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.624020 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1353568-e317-4a7d-8765-675821dac72b-serving-cert\") pod \"apiserver-7bbb656c7d-b955p\" (UID: \"a1353568-e317-4a7d-8765-675821dac72b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.624043 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-audit-dir\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.624060 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a1353568-e317-4a7d-8765-675821dac72b-etcd-client\") pod \"apiserver-7bbb656c7d-b955p\" (UID: \"a1353568-e317-4a7d-8765-675821dac72b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.624077 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bp6wb\" (UniqueName: \"kubernetes.io/projected/2f02af08-d7ad-4460-b103-cf44d7598449-kube-api-access-bp6wb\") pod \"console-operator-58897d9998-klqt2\" (UID: \"2f02af08-d7ad-4460-b103-cf44d7598449\") " pod="openshift-console-operator/console-operator-58897d9998-klqt2" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.624095 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f02af08-d7ad-4460-b103-cf44d7598449-serving-cert\") pod \"console-operator-58897d9998-klqt2\" (UID: \"2f02af08-d7ad-4460-b103-cf44d7598449\") " pod="openshift-console-operator/console-operator-58897d9998-klqt2" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.624112 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f6312e4-347f-4e82-81b9-4cb1aa530777-config\") pod \"machine-approver-56656f9798-ktnmq\" (UID: \"8f6312e4-347f-4e82-81b9-4cb1aa530777\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktnmq" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.624129 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-audit\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.624144 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-etcd-serving-ca\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.624162 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mq8d\" (UniqueName: \"kubernetes.io/projected/d57f80ec-52ac-409d-9dc1-93a89d59f714-kube-api-access-7mq8d\") pod \"authentication-operator-69f744f599-5t7m6\" (UID: \"d57f80ec-52ac-409d-9dc1-93a89d59f714\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5t7m6" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.624180 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/994c44ca-f520-4452-b267-961bd400908e-audit-policies\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.624198 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.624213 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.624242 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgmk7\" (UniqueName: \"kubernetes.io/projected/8c420468-f370-44cb-9e39-7ae8f12b461e-kube-api-access-rgmk7\") pod \"openshift-config-operator-7777fb866f-rtsv4\" (UID: \"8c420468-f370-44cb-9e39-7ae8f12b461e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rtsv4" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.624273 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/36f25c5a-3e57-4367-b306-db9661e4f7c9-service-ca\") pod \"console-f9d7485db-gb6cc\" (UID: \"36f25c5a-3e57-4367-b306-db9661e4f7c9\") " pod="openshift-console/console-f9d7485db-gb6cc" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.624324 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d5678dc-7a74-4fff-818c-e34321037686-config\") pod \"controller-manager-879f6c89f-6bbs5\" (UID: \"1d5678dc-7a74-4fff-818c-e34321037686\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6bbs5" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.624344 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d5678dc-7a74-4fff-818c-e34321037686-serving-cert\") pod \"controller-manager-879f6c89f-6bbs5\" (UID: \"1d5678dc-7a74-4fff-818c-e34321037686\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6bbs5" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.624376 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0db3a3aa-f872-43c3-ade2-58c55a448fe3-images\") pod \"machine-api-operator-5694c8668f-dq6hp\" (UID: \"0db3a3aa-f872-43c3-ade2-58c55a448fe3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dq6hp" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.624393 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/0db3a3aa-f872-43c3-ade2-58c55a448fe3-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-dq6hp\" (UID: \"0db3a3aa-f872-43c3-ade2-58c55a448fe3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dq6hp" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.624410 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b409627-4716-4bdd-a660-4f88693150a2-serving-cert\") pod \"route-controller-manager-6576b87f9c-6896q\" (UID: \"4b409627-4716-4bdd-a660-4f88693150a2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.624446 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c420468-f370-44cb-9e39-7ae8f12b461e-serving-cert\") pod \"openshift-config-operator-7777fb866f-rtsv4\" (UID: \"8c420468-f370-44cb-9e39-7ae8f12b461e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rtsv4" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.624463 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/36f25c5a-3e57-4367-b306-db9661e4f7c9-console-oauth-config\") pod \"console-f9d7485db-gb6cc\" (UID: \"36f25c5a-3e57-4367-b306-db9661e4f7c9\") " pod="openshift-console/console-f9d7485db-gb6cc" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.624578 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a1353568-e317-4a7d-8765-675821dac72b-audit-dir\") pod \"apiserver-7bbb656c7d-b955p\" (UID: \"a1353568-e317-4a7d-8765-675821dac72b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.624612 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.630846 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.632105 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.632991 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nzc7"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.633371 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nzc7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.633403 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.633520 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2twln" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.634845 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.635182 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.635227 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8c420468-f370-44cb-9e39-7ae8f12b461e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rtsv4\" (UID: \"8c420468-f370-44cb-9e39-7ae8f12b461e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rtsv4" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.635273 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-encryption-config\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.635294 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1d5678dc-7a74-4fff-818c-e34321037686-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-6bbs5\" (UID: \"1d5678dc-7a74-4fff-818c-e34321037686\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6bbs5" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.635309 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8b5mh\" (UniqueName: \"kubernetes.io/projected/994c44ca-f520-4452-b267-961bd400908e-kube-api-access-8b5mh\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.635338 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-config\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.635354 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4km4\" (UniqueName: \"kubernetes.io/projected/a1353568-e317-4a7d-8765-675821dac72b-kube-api-access-j4km4\") pod \"apiserver-7bbb656c7d-b955p\" (UID: \"a1353568-e317-4a7d-8765-675821dac72b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.635371 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d57f80ec-52ac-409d-9dc1-93a89d59f714-config\") pod \"authentication-operator-69f744f599-5t7m6\" (UID: \"d57f80ec-52ac-409d-9dc1-93a89d59f714\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5t7m6" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.635432 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8f6312e4-347f-4e82-81b9-4cb1aa530777-auth-proxy-config\") pod \"machine-approver-56656f9798-ktnmq\" (UID: \"8f6312e4-347f-4e82-81b9-4cb1aa530777\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktnmq" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.635467 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ng2h6\" (UniqueName: \"kubernetes.io/projected/0db3a3aa-f872-43c3-ade2-58c55a448fe3-kube-api-access-ng2h6\") pod \"machine-api-operator-5694c8668f-dq6hp\" (UID: \"0db3a3aa-f872-43c3-ade2-58c55a448fe3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dq6hp" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.635544 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/36f25c5a-3e57-4367-b306-db9661e4f7c9-trusted-ca-bundle\") pod \"console-f9d7485db-gb6cc\" (UID: \"36f25c5a-3e57-4367-b306-db9661e4f7c9\") " pod="openshift-console/console-f9d7485db-gb6cc" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.635568 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mm2f\" (UniqueName: \"kubernetes.io/projected/4b409627-4716-4bdd-a660-4f88693150a2-kube-api-access-7mm2f\") pod \"route-controller-manager-6576b87f9c-6896q\" (UID: \"4b409627-4716-4bdd-a660-4f88693150a2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.635582 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/36f25c5a-3e57-4367-b306-db9661e4f7c9-console-config\") pod \"console-f9d7485db-gb6cc\" (UID: \"36f25c5a-3e57-4367-b306-db9661e4f7c9\") " pod="openshift-console/console-f9d7485db-gb6cc" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.635600 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lkhl\" (UniqueName: \"kubernetes.io/projected/1d5678dc-7a74-4fff-818c-e34321037686-kube-api-access-8lkhl\") pod \"controller-manager-879f6c89f-6bbs5\" (UID: \"1d5678dc-7a74-4fff-818c-e34321037686\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6bbs5" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.635615 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/36f25c5a-3e57-4367-b306-db9661e4f7c9-console-serving-cert\") pod \"console-f9d7485db-gb6cc\" (UID: \"36f25c5a-3e57-4367-b306-db9661e4f7c9\") " pod="openshift-console/console-f9d7485db-gb6cc" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.635629 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cncsp\" (UniqueName: \"kubernetes.io/projected/36f25c5a-3e57-4367-b306-db9661e4f7c9-kube-api-access-cncsp\") pod \"console-f9d7485db-gb6cc\" (UID: \"36f25c5a-3e57-4367-b306-db9661e4f7c9\") " pod="openshift-console/console-f9d7485db-gb6cc" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.635645 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d57f80ec-52ac-409d-9dc1-93a89d59f714-service-ca-bundle\") pod \"authentication-operator-69f744f599-5t7m6\" (UID: \"d57f80ec-52ac-409d-9dc1-93a89d59f714\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5t7m6" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.635665 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.635681 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8f6312e4-347f-4e82-81b9-4cb1aa530777-machine-approver-tls\") pod \"machine-approver-56656f9798-ktnmq\" (UID: \"8f6312e4-347f-4e82-81b9-4cb1aa530777\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktnmq" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.635760 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f02af08-d7ad-4460-b103-cf44d7598449-config\") pod \"console-operator-58897d9998-klqt2\" (UID: \"2f02af08-d7ad-4460-b103-cf44d7598449\") " pod="openshift-console-operator/console-operator-58897d9998-klqt2" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.635780 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-node-pullsecrets\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.635795 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d57f80ec-52ac-409d-9dc1-93a89d59f714-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-5t7m6\" (UID: \"d57f80ec-52ac-409d-9dc1-93a89d59f714\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5t7m6" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.635806 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.635832 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62fgc\" (UniqueName: \"kubernetes.io/projected/30e8f6c3-6c7e-45dc-9a6f-e2de13d1322c-kube-api-access-62fgc\") pod \"openshift-apiserver-operator-796bbdcf4f-pmr8t\" (UID: \"30e8f6c3-6c7e-45dc-9a6f-e2de13d1322c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pmr8t" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.635856 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a1353568-e317-4a7d-8765-675821dac72b-audit-policies\") pod \"apiserver-7bbb656c7d-b955p\" (UID: \"a1353568-e317-4a7d-8765-675821dac72b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.635905 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0db3a3aa-f872-43c3-ade2-58c55a448fe3-config\") pod \"machine-api-operator-5694c8668f-dq6hp\" (UID: \"0db3a3aa-f872-43c3-ade2-58c55a448fe3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dq6hp" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.635942 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jqrv2"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.635949 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.635976 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8km7p\" (UniqueName: \"kubernetes.io/projected/8f6312e4-347f-4e82-81b9-4cb1aa530777-kube-api-access-8km7p\") pod \"machine-approver-56656f9798-ktnmq\" (UID: \"8f6312e4-347f-4e82-81b9-4cb1aa530777\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktnmq" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.636082 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.636171 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.636427 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jqrv2" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.636449 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.636646 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.641013 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.645388 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x54c"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.646135 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x54c" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.647565 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417265-hcswl"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.648085 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417265-hcswl" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.649391 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fq7x6"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.656684 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.659866 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxx9w"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.660206 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-85tkp"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.660421 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fq7x6" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.660501 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-78nhb"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.660747 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxx9w" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.660805 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-78nhb" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.660971 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-85tkp" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.661145 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-t8zwg"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.661731 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-t8zwg" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.666376 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-mwqcj"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.670044 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-289lx"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.670238 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-mwqcj" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.670329 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bzqcl"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.670997 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mllt2"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.671300 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-289lx" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.671492 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bzqcl" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.671590 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mllt2" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.674784 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-gb6cc"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.680951 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.686503 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cbbdg"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.689947 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-nvs5q"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.690303 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tw7m2"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.690647 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-rkmnv"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.691069 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cbbdg" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.691153 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tw7m2" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.691127 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nvs5q" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.691103 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rkmnv" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.696964 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-5t7m6"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.697562 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.701693 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-5cbx6"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.708353 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-ncvs7"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.709412 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-ncvs7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.710155 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5cbx6" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.715310 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pmr8t"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.716066 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-c7cx6"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.718522 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c7cx6" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.718664 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.720702 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-tm2zm"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.724126 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-wfhsc"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.724668 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-tm2zm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.724706 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-wfhsc" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.727629 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rtsv4"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.732040 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8vzrm"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.732364 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-kkbqh"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.734027 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fq7x6"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.734706 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-j79t8"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.735703 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d4h8p"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736404 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lkhl\" (UniqueName: \"kubernetes.io/projected/1d5678dc-7a74-4fff-818c-e34321037686-kube-api-access-8lkhl\") pod \"controller-manager-879f6c89f-6bbs5\" (UID: \"1d5678dc-7a74-4fff-818c-e34321037686\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6bbs5" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736430 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/36f25c5a-3e57-4367-b306-db9661e4f7c9-console-serving-cert\") pod \"console-f9d7485db-gb6cc\" (UID: \"36f25c5a-3e57-4367-b306-db9661e4f7c9\") " pod="openshift-console/console-f9d7485db-gb6cc" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736445 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cncsp\" (UniqueName: \"kubernetes.io/projected/36f25c5a-3e57-4367-b306-db9661e4f7c9-kube-api-access-cncsp\") pod \"console-f9d7485db-gb6cc\" (UID: \"36f25c5a-3e57-4367-b306-db9661e4f7c9\") " pod="openshift-console/console-f9d7485db-gb6cc" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736461 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d57f80ec-52ac-409d-9dc1-93a89d59f714-service-ca-bundle\") pod \"authentication-operator-69f744f599-5t7m6\" (UID: \"d57f80ec-52ac-409d-9dc1-93a89d59f714\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5t7m6" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736481 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736498 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f02af08-d7ad-4460-b103-cf44d7598449-config\") pod \"console-operator-58897d9998-klqt2\" (UID: \"2f02af08-d7ad-4460-b103-cf44d7598449\") " pod="openshift-console-operator/console-operator-58897d9998-klqt2" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736513 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-node-pullsecrets\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736528 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d57f80ec-52ac-409d-9dc1-93a89d59f714-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-5t7m6\" (UID: \"d57f80ec-52ac-409d-9dc1-93a89d59f714\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5t7m6" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736543 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8f6312e4-347f-4e82-81b9-4cb1aa530777-machine-approver-tls\") pod \"machine-approver-56656f9798-ktnmq\" (UID: \"8f6312e4-347f-4e82-81b9-4cb1aa530777\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktnmq" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736560 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62fgc\" (UniqueName: \"kubernetes.io/projected/30e8f6c3-6c7e-45dc-9a6f-e2de13d1322c-kube-api-access-62fgc\") pod \"openshift-apiserver-operator-796bbdcf4f-pmr8t\" (UID: \"30e8f6c3-6c7e-45dc-9a6f-e2de13d1322c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pmr8t" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736575 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a1353568-e317-4a7d-8765-675821dac72b-audit-policies\") pod \"apiserver-7bbb656c7d-b955p\" (UID: \"a1353568-e317-4a7d-8765-675821dac72b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736590 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0db3a3aa-f872-43c3-ade2-58c55a448fe3-config\") pod \"machine-api-operator-5694c8668f-dq6hp\" (UID: \"0db3a3aa-f872-43c3-ade2-58c55a448fe3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dq6hp" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736605 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736619 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8km7p\" (UniqueName: \"kubernetes.io/projected/8f6312e4-347f-4e82-81b9-4cb1aa530777-kube-api-access-8km7p\") pod \"machine-approver-56656f9798-ktnmq\" (UID: \"8f6312e4-347f-4e82-81b9-4cb1aa530777\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktnmq" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736633 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/30e8f6c3-6c7e-45dc-9a6f-e2de13d1322c-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-pmr8t\" (UID: \"30e8f6c3-6c7e-45dc-9a6f-e2de13d1322c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pmr8t" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736648 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736662 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d57f80ec-52ac-409d-9dc1-93a89d59f714-serving-cert\") pod \"authentication-operator-69f744f599-5t7m6\" (UID: \"d57f80ec-52ac-409d-9dc1-93a89d59f714\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5t7m6" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736676 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/994c44ca-f520-4452-b267-961bd400908e-audit-dir\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736690 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5qmj\" (UniqueName: \"kubernetes.io/projected/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-kube-api-access-l5qmj\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736712 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4b409627-4716-4bdd-a660-4f88693150a2-client-ca\") pod \"route-controller-manager-6576b87f9c-6896q\" (UID: \"4b409627-4716-4bdd-a660-4f88693150a2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736728 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a1353568-e317-4a7d-8765-675821dac72b-encryption-config\") pod \"apiserver-7bbb656c7d-b955p\" (UID: \"a1353568-e317-4a7d-8765-675821dac72b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736742 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-etcd-client\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736772 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-serving-cert\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736788 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736806 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b409627-4716-4bdd-a660-4f88693150a2-config\") pod \"route-controller-manager-6576b87f9c-6896q\" (UID: \"4b409627-4716-4bdd-a660-4f88693150a2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736822 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736839 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a1353568-e317-4a7d-8765-675821dac72b-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-b955p\" (UID: \"a1353568-e317-4a7d-8765-675821dac72b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736853 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2f02af08-d7ad-4460-b103-cf44d7598449-trusted-ca\") pod \"console-operator-58897d9998-klqt2\" (UID: \"2f02af08-d7ad-4460-b103-cf44d7598449\") " pod="openshift-console-operator/console-operator-58897d9998-klqt2" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736867 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-image-import-ca\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736883 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a1353568-e317-4a7d-8765-675821dac72b-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-b955p\" (UID: \"a1353568-e317-4a7d-8765-675821dac72b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736897 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30e8f6c3-6c7e-45dc-9a6f-e2de13d1322c-config\") pod \"openshift-apiserver-operator-796bbdcf4f-pmr8t\" (UID: \"30e8f6c3-6c7e-45dc-9a6f-e2de13d1322c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pmr8t" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736915 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736930 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1d5678dc-7a74-4fff-818c-e34321037686-client-ca\") pod \"controller-manager-879f6c89f-6bbs5\" (UID: \"1d5678dc-7a74-4fff-818c-e34321037686\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6bbs5" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736945 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736960 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/36f25c5a-3e57-4367-b306-db9661e4f7c9-oauth-serving-cert\") pod \"console-f9d7485db-gb6cc\" (UID: \"36f25c5a-3e57-4367-b306-db9661e4f7c9\") " pod="openshift-console/console-f9d7485db-gb6cc" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736975 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.736993 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1353568-e317-4a7d-8765-675821dac72b-serving-cert\") pod \"apiserver-7bbb656c7d-b955p\" (UID: \"a1353568-e317-4a7d-8765-675821dac72b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737008 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-audit-dir\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737022 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a1353568-e317-4a7d-8765-675821dac72b-etcd-client\") pod \"apiserver-7bbb656c7d-b955p\" (UID: \"a1353568-e317-4a7d-8765-675821dac72b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737037 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bp6wb\" (UniqueName: \"kubernetes.io/projected/2f02af08-d7ad-4460-b103-cf44d7598449-kube-api-access-bp6wb\") pod \"console-operator-58897d9998-klqt2\" (UID: \"2f02af08-d7ad-4460-b103-cf44d7598449\") " pod="openshift-console-operator/console-operator-58897d9998-klqt2" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737053 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f02af08-d7ad-4460-b103-cf44d7598449-serving-cert\") pod \"console-operator-58897d9998-klqt2\" (UID: \"2f02af08-d7ad-4460-b103-cf44d7598449\") " pod="openshift-console-operator/console-operator-58897d9998-klqt2" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737067 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f6312e4-347f-4e82-81b9-4cb1aa530777-config\") pod \"machine-approver-56656f9798-ktnmq\" (UID: \"8f6312e4-347f-4e82-81b9-4cb1aa530777\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktnmq" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737081 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-audit\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737096 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mq8d\" (UniqueName: \"kubernetes.io/projected/d57f80ec-52ac-409d-9dc1-93a89d59f714-kube-api-access-7mq8d\") pod \"authentication-operator-69f744f599-5t7m6\" (UID: \"d57f80ec-52ac-409d-9dc1-93a89d59f714\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5t7m6" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737112 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/994c44ca-f520-4452-b267-961bd400908e-audit-policies\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737127 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737143 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737159 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-etcd-serving-ca\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737181 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgmk7\" (UniqueName: \"kubernetes.io/projected/8c420468-f370-44cb-9e39-7ae8f12b461e-kube-api-access-rgmk7\") pod \"openshift-config-operator-7777fb866f-rtsv4\" (UID: \"8c420468-f370-44cb-9e39-7ae8f12b461e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rtsv4" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737195 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/36f25c5a-3e57-4367-b306-db9661e4f7c9-service-ca\") pod \"console-f9d7485db-gb6cc\" (UID: \"36f25c5a-3e57-4367-b306-db9661e4f7c9\") " pod="openshift-console/console-f9d7485db-gb6cc" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737212 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d5678dc-7a74-4fff-818c-e34321037686-config\") pod \"controller-manager-879f6c89f-6bbs5\" (UID: \"1d5678dc-7a74-4fff-818c-e34321037686\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6bbs5" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737226 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d5678dc-7a74-4fff-818c-e34321037686-serving-cert\") pod \"controller-manager-879f6c89f-6bbs5\" (UID: \"1d5678dc-7a74-4fff-818c-e34321037686\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6bbs5" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737240 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0db3a3aa-f872-43c3-ade2-58c55a448fe3-images\") pod \"machine-api-operator-5694c8668f-dq6hp\" (UID: \"0db3a3aa-f872-43c3-ade2-58c55a448fe3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dq6hp" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737267 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/0db3a3aa-f872-43c3-ade2-58c55a448fe3-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-dq6hp\" (UID: \"0db3a3aa-f872-43c3-ade2-58c55a448fe3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dq6hp" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737283 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b409627-4716-4bdd-a660-4f88693150a2-serving-cert\") pod \"route-controller-manager-6576b87f9c-6896q\" (UID: \"4b409627-4716-4bdd-a660-4f88693150a2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737308 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c420468-f370-44cb-9e39-7ae8f12b461e-serving-cert\") pod \"openshift-config-operator-7777fb866f-rtsv4\" (UID: \"8c420468-f370-44cb-9e39-7ae8f12b461e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rtsv4" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737322 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a1353568-e317-4a7d-8765-675821dac72b-audit-dir\") pod \"apiserver-7bbb656c7d-b955p\" (UID: \"a1353568-e317-4a7d-8765-675821dac72b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737343 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737357 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737373 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8c420468-f370-44cb-9e39-7ae8f12b461e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rtsv4\" (UID: \"8c420468-f370-44cb-9e39-7ae8f12b461e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rtsv4" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737388 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/36f25c5a-3e57-4367-b306-db9661e4f7c9-console-oauth-config\") pod \"console-f9d7485db-gb6cc\" (UID: \"36f25c5a-3e57-4367-b306-db9661e4f7c9\") " pod="openshift-console/console-f9d7485db-gb6cc" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737402 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-encryption-config\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737415 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1d5678dc-7a74-4fff-818c-e34321037686-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-6bbs5\" (UID: \"1d5678dc-7a74-4fff-818c-e34321037686\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6bbs5" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737430 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8b5mh\" (UniqueName: \"kubernetes.io/projected/994c44ca-f520-4452-b267-961bd400908e-kube-api-access-8b5mh\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737444 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-config\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737458 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4km4\" (UniqueName: \"kubernetes.io/projected/a1353568-e317-4a7d-8765-675821dac72b-kube-api-access-j4km4\") pod \"apiserver-7bbb656c7d-b955p\" (UID: \"a1353568-e317-4a7d-8765-675821dac72b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737472 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d57f80ec-52ac-409d-9dc1-93a89d59f714-config\") pod \"authentication-operator-69f744f599-5t7m6\" (UID: \"d57f80ec-52ac-409d-9dc1-93a89d59f714\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5t7m6" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737486 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8f6312e4-347f-4e82-81b9-4cb1aa530777-auth-proxy-config\") pod \"machine-approver-56656f9798-ktnmq\" (UID: \"8f6312e4-347f-4e82-81b9-4cb1aa530777\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktnmq" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737501 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ng2h6\" (UniqueName: \"kubernetes.io/projected/0db3a3aa-f872-43c3-ade2-58c55a448fe3-kube-api-access-ng2h6\") pod \"machine-api-operator-5694c8668f-dq6hp\" (UID: \"0db3a3aa-f872-43c3-ade2-58c55a448fe3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dq6hp" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737524 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mm2f\" (UniqueName: \"kubernetes.io/projected/4b409627-4716-4bdd-a660-4f88693150a2-kube-api-access-7mm2f\") pod \"route-controller-manager-6576b87f9c-6896q\" (UID: \"4b409627-4716-4bdd-a660-4f88693150a2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737539 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/36f25c5a-3e57-4367-b306-db9661e4f7c9-console-config\") pod \"console-f9d7485db-gb6cc\" (UID: \"36f25c5a-3e57-4367-b306-db9661e4f7c9\") " pod="openshift-console/console-f9d7485db-gb6cc" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.737573 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/36f25c5a-3e57-4367-b306-db9661e4f7c9-trusted-ca-bundle\") pod \"console-f9d7485db-gb6cc\" (UID: \"36f25c5a-3e57-4367-b306-db9661e4f7c9\") " pod="openshift-console/console-f9d7485db-gb6cc" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.738439 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.738462 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/36f25c5a-3e57-4367-b306-db9661e4f7c9-trusted-ca-bundle\") pod \"console-f9d7485db-gb6cc\" (UID: \"36f25c5a-3e57-4367-b306-db9661e4f7c9\") " pod="openshift-console/console-f9d7485db-gb6cc" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.738681 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-85tkp"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.739802 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x54c"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.740241 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d57f80ec-52ac-409d-9dc1-93a89d59f714-service-ca-bundle\") pod \"authentication-operator-69f744f599-5t7m6\" (UID: \"d57f80ec-52ac-409d-9dc1-93a89d59f714\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5t7m6" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.740425 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/994c44ca-f520-4452-b267-961bd400908e-audit-policies\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.740487 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-audit-dir\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.740545 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1d5678dc-7a74-4fff-818c-e34321037686-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-6bbs5\" (UID: \"1d5678dc-7a74-4fff-818c-e34321037686\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6bbs5" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.741066 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.741701 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f02af08-d7ad-4460-b103-cf44d7598449-config\") pod \"console-operator-58897d9998-klqt2\" (UID: \"2f02af08-d7ad-4460-b103-cf44d7598449\") " pod="openshift-console-operator/console-operator-58897d9998-klqt2" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.741752 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-node-pullsecrets\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.742880 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d57f80ec-52ac-409d-9dc1-93a89d59f714-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-5t7m6\" (UID: \"d57f80ec-52ac-409d-9dc1-93a89d59f714\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5t7m6" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.745012 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b409627-4716-4bdd-a660-4f88693150a2-serving-cert\") pod \"route-controller-manager-6576b87f9c-6896q\" (UID: \"4b409627-4716-4bdd-a660-4f88693150a2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.745248 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8f6312e4-347f-4e82-81b9-4cb1aa530777-machine-approver-tls\") pod \"machine-approver-56656f9798-ktnmq\" (UID: \"8f6312e4-347f-4e82-81b9-4cb1aa530777\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktnmq" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.745890 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a1353568-e317-4a7d-8765-675821dac72b-audit-dir\") pod \"apiserver-7bbb656c7d-b955p\" (UID: \"a1353568-e317-4a7d-8765-675821dac72b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.747791 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nzc7"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.749474 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8c420468-f370-44cb-9e39-7ae8f12b461e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rtsv4\" (UID: \"8c420468-f370-44cb-9e39-7ae8f12b461e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rtsv4" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.749544 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a1353568-e317-4a7d-8765-675821dac72b-audit-policies\") pod \"apiserver-7bbb656c7d-b955p\" (UID: \"a1353568-e317-4a7d-8765-675821dac72b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.750019 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/36f25c5a-3e57-4367-b306-db9661e4f7c9-service-ca\") pod \"console-f9d7485db-gb6cc\" (UID: \"36f25c5a-3e57-4367-b306-db9661e4f7c9\") " pod="openshift-console/console-f9d7485db-gb6cc" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.751160 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/994c44ca-f520-4452-b267-961bd400908e-audit-dir\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.751660 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d5678dc-7a74-4fff-818c-e34321037686-config\") pod \"controller-manager-879f6c89f-6bbs5\" (UID: \"1d5678dc-7a74-4fff-818c-e34321037686\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6bbs5" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.752309 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jqrv2"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.752334 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-mwqcj"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.752345 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-dld8t"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.752492 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0db3a3aa-f872-43c3-ade2-58c55a448fe3-images\") pod \"machine-api-operator-5694c8668f-dq6hp\" (UID: \"0db3a3aa-f872-43c3-ade2-58c55a448fe3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dq6hp" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.752884 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.752985 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-etcd-serving-ca\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.753364 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.753707 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0db3a3aa-f872-43c3-ade2-58c55a448fe3-config\") pod \"machine-api-operator-5694c8668f-dq6hp\" (UID: \"0db3a3aa-f872-43c3-ade2-58c55a448fe3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dq6hp" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.753720 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.754164 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d57f80ec-52ac-409d-9dc1-93a89d59f714-config\") pod \"authentication-operator-69f744f599-5t7m6\" (UID: \"d57f80ec-52ac-409d-9dc1-93a89d59f714\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5t7m6" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.752955 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-dld8t" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.755448 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-config\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.755151 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c420468-f370-44cb-9e39-7ae8f12b461e-serving-cert\") pod \"openshift-config-operator-7777fb866f-rtsv4\" (UID: \"8c420468-f370-44cb-9e39-7ae8f12b461e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rtsv4" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.755970 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-audit\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.756228 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/36f25c5a-3e57-4367-b306-db9661e4f7c9-console-config\") pod \"console-f9d7485db-gb6cc\" (UID: \"36f25c5a-3e57-4367-b306-db9661e4f7c9\") " pod="openshift-console/console-f9d7485db-gb6cc" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.756321 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/36f25c5a-3e57-4367-b306-db9661e4f7c9-oauth-serving-cert\") pod \"console-f9d7485db-gb6cc\" (UID: \"36f25c5a-3e57-4367-b306-db9661e4f7c9\") " pod="openshift-console/console-f9d7485db-gb6cc" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.756575 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2f02af08-d7ad-4460-b103-cf44d7598449-trusted-ca\") pod \"console-operator-58897d9998-klqt2\" (UID: \"2f02af08-d7ad-4460-b103-cf44d7598449\") " pod="openshift-console-operator/console-operator-58897d9998-klqt2" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.756622 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4b409627-4716-4bdd-a660-4f88693150a2-client-ca\") pod \"route-controller-manager-6576b87f9c-6896q\" (UID: \"4b409627-4716-4bdd-a660-4f88693150a2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.756924 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f6312e4-347f-4e82-81b9-4cb1aa530777-config\") pod \"machine-approver-56656f9798-ktnmq\" (UID: \"8f6312e4-347f-4e82-81b9-4cb1aa530777\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktnmq" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.757211 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.757395 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8f6312e4-347f-4e82-81b9-4cb1aa530777-auth-proxy-config\") pod \"machine-approver-56656f9798-ktnmq\" (UID: \"8f6312e4-347f-4e82-81b9-4cb1aa530777\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktnmq" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.758110 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a1353568-e317-4a7d-8765-675821dac72b-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-b955p\" (UID: \"a1353568-e317-4a7d-8765-675821dac72b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.758210 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-image-import-ca\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.755054 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-kz9wk"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.758654 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1d5678dc-7a74-4fff-818c-e34321037686-client-ca\") pod \"controller-manager-879f6c89f-6bbs5\" (UID: \"1d5678dc-7a74-4fff-818c-e34321037686\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6bbs5" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.758801 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.759272 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a1353568-e317-4a7d-8765-675821dac72b-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-b955p\" (UID: \"a1353568-e317-4a7d-8765-675821dac72b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.759335 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-encryption-config\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.759468 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.759700 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.759958 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cbbdg"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.759985 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tw7m2"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.759986 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b409627-4716-4bdd-a660-4f88693150a2-config\") pod \"route-controller-manager-6576b87f9c-6896q\" (UID: \"4b409627-4716-4bdd-a660-4f88693150a2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.760048 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30e8f6c3-6c7e-45dc-9a6f-e2de13d1322c-config\") pod \"openshift-apiserver-operator-796bbdcf4f-pmr8t\" (UID: \"30e8f6c3-6c7e-45dc-9a6f-e2de13d1322c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pmr8t" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.760059 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-kz9wk" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.760237 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f02af08-d7ad-4460-b103-cf44d7598449-serving-cert\") pod \"console-operator-58897d9998-klqt2\" (UID: \"2f02af08-d7ad-4460-b103-cf44d7598449\") " pod="openshift-console-operator/console-operator-58897d9998-klqt2" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.759997 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d57f80ec-52ac-409d-9dc1-93a89d59f714-serving-cert\") pod \"authentication-operator-69f744f599-5t7m6\" (UID: \"d57f80ec-52ac-409d-9dc1-93a89d59f714\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5t7m6" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.760561 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d5678dc-7a74-4fff-818c-e34321037686-serving-cert\") pod \"controller-manager-879f6c89f-6bbs5\" (UID: \"1d5678dc-7a74-4fff-818c-e34321037686\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6bbs5" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.761456 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a1353568-e317-4a7d-8765-675821dac72b-etcd-client\") pod \"apiserver-7bbb656c7d-b955p\" (UID: \"a1353568-e317-4a7d-8765-675821dac72b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.761665 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/30e8f6c3-6c7e-45dc-9a6f-e2de13d1322c-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-pmr8t\" (UID: \"30e8f6c3-6c7e-45dc-9a6f-e2de13d1322c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pmr8t" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.761916 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/0db3a3aa-f872-43c3-ade2-58c55a448fe3-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-dq6hp\" (UID: \"0db3a3aa-f872-43c3-ade2-58c55a448fe3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dq6hp" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.761952 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.762365 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1353568-e317-4a7d-8765-675821dac72b-serving-cert\") pod \"apiserver-7bbb656c7d-b955p\" (UID: \"a1353568-e317-4a7d-8765-675821dac72b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.762699 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-etcd-client\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.762877 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-serving-cert\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.763250 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/36f25c5a-3e57-4367-b306-db9661e4f7c9-console-oauth-config\") pod \"console-f9d7485db-gb6cc\" (UID: \"36f25c5a-3e57-4367-b306-db9661e4f7c9\") " pod="openshift-console/console-f9d7485db-gb6cc" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.763521 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.764027 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/36f25c5a-3e57-4367-b306-db9661e4f7c9-console-serving-cert\") pod \"console-f9d7485db-gb6cc\" (UID: \"36f25c5a-3e57-4367-b306-db9661e4f7c9\") " pod="openshift-console/console-f9d7485db-gb6cc" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.764234 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a1353568-e317-4a7d-8765-675821dac72b-encryption-config\") pod \"apiserver-7bbb656c7d-b955p\" (UID: \"a1353568-e317-4a7d-8765-675821dac72b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.764657 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.765737 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-nvs5q"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.767018 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-c7cx6"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.768281 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzb98"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.769670 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-78nhb"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.771108 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.772014 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxx9w"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.773070 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-2twln"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.774136 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bzqcl"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.775300 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-289lx"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.776559 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-tm2zm"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.777086 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.778110 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-rkmnv"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.782514 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417265-hcswl"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.784487 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-5cbx6"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.788615 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-t8zwg"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.790499 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mllt2"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.792550 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-dld8t"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.794543 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-wfhsc"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.796738 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.799025 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-wj67k"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.800228 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-wj67k" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.800375 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-wj67k"] Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.816522 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.877184 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.897148 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.917738 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.937685 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.957250 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.977104 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 15:48:18 crc kubenswrapper[4813]: I1206 15:48:18.997896 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.017790 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.038378 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.058630 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.077156 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.098143 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.117646 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.138198 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.157371 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.177775 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.197592 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.218713 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.237630 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.258387 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.278774 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.297988 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.318192 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.337837 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.357830 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.377677 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.398578 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.417610 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.438559 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.458723 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.477418 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.498321 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.529308 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.538143 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.557669 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.578302 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.597733 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.618194 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.637771 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.657393 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.677890 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.695886 4813 request.go:700] Waited for 1.004097297s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator-operator/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.698211 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.717923 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.738664 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 06 15:48:19 crc kubenswrapper[4813]: E1206 15:48:19.747946 4813 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-service-ca: failed to sync configmap cache: timed out waiting for the condition Dec 06 15:48:19 crc kubenswrapper[4813]: E1206 15:48:19.748054 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-service-ca podName:994c44ca-f520-4452-b267-961bd400908e nodeName:}" failed. No retries permitted until 2025-12-06 15:48:20.2480242 +0000 UTC m=+140.138903806 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-service-ca" (UniqueName: "kubernetes.io/configmap/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-service-ca") pod "oauth-openshift-558db77b4-fszk7" (UID: "994c44ca-f520-4452-b267-961bd400908e") : failed to sync configmap cache: timed out waiting for the condition Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.758153 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.778211 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.797173 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.818657 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.837990 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.857354 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.878158 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.897799 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.917378 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.938345 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.957507 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.978381 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 06 15:48:19 crc kubenswrapper[4813]: I1206 15:48:19.998821 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.020129 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.039551 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.058167 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.078111 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.112156 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.118594 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.137668 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.157389 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.178401 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.197763 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.217910 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.237885 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.256585 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.258500 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.277245 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.298568 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.318139 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.338810 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.382198 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lkhl\" (UniqueName: \"kubernetes.io/projected/1d5678dc-7a74-4fff-818c-e34321037686-kube-api-access-8lkhl\") pod \"controller-manager-879f6c89f-6bbs5\" (UID: \"1d5678dc-7a74-4fff-818c-e34321037686\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6bbs5" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.405235 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5qmj\" (UniqueName: \"kubernetes.io/projected/4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a-kube-api-access-l5qmj\") pod \"apiserver-76f77b778f-8vzrm\" (UID: \"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a\") " pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.438848 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cncsp\" (UniqueName: \"kubernetes.io/projected/36f25c5a-3e57-4367-b306-db9661e4f7c9-kube-api-access-cncsp\") pod \"console-f9d7485db-gb6cc\" (UID: \"36f25c5a-3e57-4367-b306-db9661e4f7c9\") " pod="openshift-console/console-f9d7485db-gb6cc" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.444494 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62fgc\" (UniqueName: \"kubernetes.io/projected/30e8f6c3-6c7e-45dc-9a6f-e2de13d1322c-kube-api-access-62fgc\") pod \"openshift-apiserver-operator-796bbdcf4f-pmr8t\" (UID: \"30e8f6c3-6c7e-45dc-9a6f-e2de13d1322c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pmr8t" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.470081 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bp6wb\" (UniqueName: \"kubernetes.io/projected/2f02af08-d7ad-4460-b103-cf44d7598449-kube-api-access-bp6wb\") pod \"console-operator-58897d9998-klqt2\" (UID: \"2f02af08-d7ad-4460-b103-cf44d7598449\") " pod="openshift-console-operator/console-operator-58897d9998-klqt2" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.481823 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.489733 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8km7p\" (UniqueName: \"kubernetes.io/projected/8f6312e4-347f-4e82-81b9-4cb1aa530777-kube-api-access-8km7p\") pod \"machine-approver-56656f9798-ktnmq\" (UID: \"8f6312e4-347f-4e82-81b9-4cb1aa530777\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktnmq" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.491875 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pmr8t" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.510309 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-gb6cc" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.515009 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgmk7\" (UniqueName: \"kubernetes.io/projected/8c420468-f370-44cb-9e39-7ae8f12b461e-kube-api-access-rgmk7\") pod \"openshift-config-operator-7777fb866f-rtsv4\" (UID: \"8c420468-f370-44cb-9e39-7ae8f12b461e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rtsv4" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.522990 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ng2h6\" (UniqueName: \"kubernetes.io/projected/0db3a3aa-f872-43c3-ade2-58c55a448fe3-kube-api-access-ng2h6\") pod \"machine-api-operator-5694c8668f-dq6hp\" (UID: \"0db3a3aa-f872-43c3-ade2-58c55a448fe3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dq6hp" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.534195 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8b5mh\" (UniqueName: \"kubernetes.io/projected/994c44ca-f520-4452-b267-961bd400908e-kube-api-access-8b5mh\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.558760 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mm2f\" (UniqueName: \"kubernetes.io/projected/4b409627-4716-4bdd-a660-4f88693150a2-kube-api-access-7mm2f\") pod \"route-controller-manager-6576b87f9c-6896q\" (UID: \"4b409627-4716-4bdd-a660-4f88693150a2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.574525 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-6bbs5" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.579050 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4km4\" (UniqueName: \"kubernetes.io/projected/a1353568-e317-4a7d-8765-675821dac72b-kube-api-access-j4km4\") pod \"apiserver-7bbb656c7d-b955p\" (UID: \"a1353568-e317-4a7d-8765-675821dac72b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.597493 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mq8d\" (UniqueName: \"kubernetes.io/projected/d57f80ec-52ac-409d-9dc1-93a89d59f714-kube-api-access-7mq8d\") pod \"authentication-operator-69f744f599-5t7m6\" (UID: \"d57f80ec-52ac-409d-9dc1-93a89d59f714\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5t7m6" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.597758 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.619746 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.620161 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.636508 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-dq6hp" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.638421 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.658430 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.678286 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.696094 4813 request.go:700] Waited for 1.935727625s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmachine-config-server-tls&limit=500&resourceVersion=0 Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.697495 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.725363 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.733625 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.736065 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-gb6cc"] Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.740689 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.748510 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-klqt2" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.757845 4813 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.769294 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktnmq" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.776868 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.776892 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-5t7m6" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.802465 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rtsv4" Dec 06 15:48:20 crc kubenswrapper[4813]: W1206 15:48:20.818680 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f6312e4_347f_4e82_81b9_4cb1aa530777.slice/crio-94906f6a1745391193e48f94cb35d2b28daa01c982b6dca9d676ee0a780ee634 WatchSource:0}: Error finding container 94906f6a1745391193e48f94cb35d2b28daa01c982b6dca9d676ee0a780ee634: Status 404 returned error can't find the container with id 94906f6a1745391193e48f94cb35d2b28daa01c982b6dca9d676ee0a780ee634 Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.838564 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.849493 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-fszk7\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.859163 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6bbs5"] Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.863703 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4674583b-142e-4968-9824-7090cfebca75-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lzb98\" (UID: \"4674583b-142e-4968-9824-7090cfebca75\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzb98" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.863743 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d2857e44-0b98-4f13-baca-3330474d8b44-bound-sa-token\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.863865 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a7d392b-a4af-4e95-b50e-3e45d8c01037-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-6nzc7\" (UID: \"8a7d392b-a4af-4e95-b50e-3e45d8c01037\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nzc7" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.863903 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d2857e44-0b98-4f13-baca-3330474d8b44-registry-tls\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.864123 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.864154 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/4674583b-142e-4968-9824-7090cfebca75-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lzb98\" (UID: \"4674583b-142e-4968-9824-7090cfebca75\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzb98" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.864181 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1883cf65-656c-439d-980d-597c40899cc5-proxy-tls\") pod \"machine-config-operator-74547568cd-2twln\" (UID: \"1883cf65-656c-439d-980d-597c40899cc5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2twln" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.864249 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c5ca581c-83f3-472a-962e-2edaf6e3eba9-tmpfs\") pod \"packageserver-d55dfcdfc-jqrv2\" (UID: \"c5ca581c-83f3-472a-962e-2edaf6e3eba9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jqrv2" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.864299 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8x7k9\" (UniqueName: \"kubernetes.io/projected/8a7d392b-a4af-4e95-b50e-3e45d8c01037-kube-api-access-8x7k9\") pod \"openshift-controller-manager-operator-756b6f6bc6-6nzc7\" (UID: \"8a7d392b-a4af-4e95-b50e-3e45d8c01037\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nzc7" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.864343 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2bzg\" (UniqueName: \"kubernetes.io/projected/9549da48-ad83-4eff-878d-c9df63e66cd7-kube-api-access-n2bzg\") pod \"cluster-samples-operator-665b6dd947-d4h8p\" (UID: \"9549da48-ad83-4eff-878d-c9df63e66cd7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d4h8p" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.864384 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d2857e44-0b98-4f13-baca-3330474d8b44-ca-trust-extracted\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.864411 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d2857e44-0b98-4f13-baca-3330474d8b44-trusted-ca\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.864428 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67mng\" (UniqueName: \"kubernetes.io/projected/d2857e44-0b98-4f13-baca-3330474d8b44-kube-api-access-67mng\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.864448 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9549da48-ad83-4eff-878d-c9df63e66cd7-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-d4h8p\" (UID: \"9549da48-ad83-4eff-878d-c9df63e66cd7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d4h8p" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.864506 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d2857e44-0b98-4f13-baca-3330474d8b44-registry-certificates\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.864533 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d2857e44-0b98-4f13-baca-3330474d8b44-installation-pull-secrets\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.864548 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c5ca581c-83f3-472a-962e-2edaf6e3eba9-apiservice-cert\") pod \"packageserver-d55dfcdfc-jqrv2\" (UID: \"c5ca581c-83f3-472a-962e-2edaf6e3eba9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jqrv2" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.864638 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1883cf65-656c-439d-980d-597c40899cc5-images\") pod \"machine-config-operator-74547568cd-2twln\" (UID: \"1883cf65-656c-439d-980d-597c40899cc5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2twln" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.864656 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rhp6\" (UniqueName: \"kubernetes.io/projected/1883cf65-656c-439d-980d-597c40899cc5-kube-api-access-7rhp6\") pod \"machine-config-operator-74547568cd-2twln\" (UID: \"1883cf65-656c-439d-980d-597c40899cc5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2twln" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.864690 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxb62\" (UniqueName: \"kubernetes.io/projected/4674583b-142e-4968-9824-7090cfebca75-kube-api-access-jxb62\") pod \"cluster-image-registry-operator-dc59b4c8b-lzb98\" (UID: \"4674583b-142e-4968-9824-7090cfebca75\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzb98" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.864706 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4674583b-142e-4968-9824-7090cfebca75-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lzb98\" (UID: \"4674583b-142e-4968-9824-7090cfebca75\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzb98" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.864733 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a7d392b-a4af-4e95-b50e-3e45d8c01037-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-6nzc7\" (UID: \"8a7d392b-a4af-4e95-b50e-3e45d8c01037\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nzc7" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.864748 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c5ca581c-83f3-472a-962e-2edaf6e3eba9-webhook-cert\") pod \"packageserver-d55dfcdfc-jqrv2\" (UID: \"c5ca581c-83f3-472a-962e-2edaf6e3eba9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jqrv2" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.864823 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7tbr\" (UniqueName: \"kubernetes.io/projected/c5ca581c-83f3-472a-962e-2edaf6e3eba9-kube-api-access-p7tbr\") pod \"packageserver-d55dfcdfc-jqrv2\" (UID: \"c5ca581c-83f3-472a-962e-2edaf6e3eba9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jqrv2" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.864879 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1883cf65-656c-439d-980d-597c40899cc5-auth-proxy-config\") pod \"machine-config-operator-74547568cd-2twln\" (UID: \"1883cf65-656c-439d-980d-597c40899cc5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2twln" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.864897 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjpjn\" (UniqueName: \"kubernetes.io/projected/26bc2fa2-9eee-4f7b-b31b-86f364df2b06-kube-api-access-zjpjn\") pod \"downloads-7954f5f757-kkbqh\" (UID: \"26bc2fa2-9eee-4f7b-b31b-86f364df2b06\") " pod="openshift-console/downloads-7954f5f757-kkbqh" Dec 06 15:48:20 crc kubenswrapper[4813]: E1206 15:48:20.865178 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:21.365165268 +0000 UTC m=+141.256044844 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.893533 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-gb6cc" event={"ID":"36f25c5a-3e57-4367-b306-db9661e4f7c9","Type":"ContainerStarted","Data":"299d968a296378c9098c081177ae4a317e4111ddf0168a491ed8eefde4e399e6"} Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.895003 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktnmq" event={"ID":"8f6312e4-347f-4e82-81b9-4cb1aa530777","Type":"ContainerStarted","Data":"94906f6a1745391193e48f94cb35d2b28daa01c982b6dca9d676ee0a780ee634"} Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.905360 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-6bbs5" event={"ID":"1d5678dc-7a74-4fff-818c-e34321037686","Type":"ContainerStarted","Data":"beffd2960d07cdfd8fb916176928a691fd0c55570d38ae8bb6ef245ce63da5cc"} Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.967363 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.967514 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82cp2\" (UniqueName: \"kubernetes.io/projected/2e8f4e1f-b111-4e8f-8600-6ac82ded86e9-kube-api-access-82cp2\") pod \"multus-admission-controller-857f4d67dd-t8zwg\" (UID: \"2e8f4e1f-b111-4e8f-8600-6ac82ded86e9\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t8zwg" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.967545 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2bzg\" (UniqueName: \"kubernetes.io/projected/9549da48-ad83-4eff-878d-c9df63e66cd7-kube-api-access-n2bzg\") pod \"cluster-samples-operator-665b6dd947-d4h8p\" (UID: \"9549da48-ad83-4eff-878d-c9df63e66cd7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d4h8p" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.967566 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2e8f4e1f-b111-4e8f-8600-6ac82ded86e9-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-t8zwg\" (UID: \"2e8f4e1f-b111-4e8f-8600-6ac82ded86e9\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t8zwg" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.967581 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8ddf8368-8232-4211-bfca-642b1acef6a4-default-certificate\") pod \"router-default-5444994796-ncvs7\" (UID: \"8ddf8368-8232-4211-bfca-642b1acef6a4\") " pod="openshift-ingress/router-default-5444994796-ncvs7" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.967596 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0664e3bf-bbbe-4547-9697-785efea4dd99-trusted-ca\") pod \"ingress-operator-5b745b69d9-5cbx6\" (UID: \"0664e3bf-bbbe-4547-9697-785efea4dd99\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5cbx6" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.967629 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d2857e44-0b98-4f13-baca-3330474d8b44-ca-trust-extracted\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.967655 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1e457a6a-af4c-4d2b-a2af-c758de365a40-metrics-tls\") pod \"dns-default-tm2zm\" (UID: \"1e457a6a-af4c-4d2b-a2af-c758de365a40\") " pod="openshift-dns/dns-default-tm2zm" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.967670 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/45d5fd66-73f2-4eaa-88ad-4cc6bc597387-signing-cabundle\") pod \"service-ca-9c57cc56f-wfhsc\" (UID: \"45d5fd66-73f2-4eaa-88ad-4cc6bc597387\") " pod="openshift-service-ca/service-ca-9c57cc56f-wfhsc" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.967688 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb28f2df-5922-47fc-aaf0-5887ca9eb03c-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9x54c\" (UID: \"bb28f2df-5922-47fc-aaf0-5887ca9eb03c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x54c" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.967704 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b0e59fb-3d78-4720-bcc2-2c2eb985b7bf-config\") pod \"service-ca-operator-777779d784-nvs5q\" (UID: \"8b0e59fb-3d78-4720-bcc2-2c2eb985b7bf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nvs5q" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.967718 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36af9565-0192-4fee-9729-2bb06631e150-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-cbbdg\" (UID: \"36af9565-0192-4fee-9729-2bb06631e150\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cbbdg" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.967735 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6687x\" (UniqueName: \"kubernetes.io/projected/935ac385-32ee-46c5-bffb-69b89b4af6a9-kube-api-access-6687x\") pod \"control-plane-machine-set-operator-78cbb6b69f-mllt2\" (UID: \"935ac385-32ee-46c5-bffb-69b89b4af6a9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mllt2" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.967748 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb3d374a-97d3-4cf8-972a-8e8b795eb9b1-serving-cert\") pod \"etcd-operator-b45778765-78nhb\" (UID: \"bb3d374a-97d3-4cf8-972a-8e8b795eb9b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-78nhb" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.967763 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ddf8368-8232-4211-bfca-642b1acef6a4-service-ca-bundle\") pod \"router-default-5444994796-ncvs7\" (UID: \"8ddf8368-8232-4211-bfca-642b1acef6a4\") " pod="openshift-ingress/router-default-5444994796-ncvs7" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.967779 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a238598-92cd-4641-ae82-6e471740ee7f-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-85tkp\" (UID: \"9a238598-92cd-4641-ae82-6e471740ee7f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-85tkp" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.967792 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/279e4882-ef42-445a-891f-65dff6bdd8dc-cert\") pod \"ingress-canary-dld8t\" (UID: \"279e4882-ef42-445a-891f-65dff6bdd8dc\") " pod="openshift-ingress-canary/ingress-canary-dld8t" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.967806 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e8aef227-c82b-4e7d-bf13-d99c935ad12b-profile-collector-cert\") pod \"catalog-operator-68c6474976-bzqcl\" (UID: \"e8aef227-c82b-4e7d-bf13-d99c935ad12b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bzqcl" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.967825 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1883cf65-656c-439d-980d-597c40899cc5-images\") pod \"machine-config-operator-74547568cd-2twln\" (UID: \"1883cf65-656c-439d-980d-597c40899cc5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2twln" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.967839 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bb3d374a-97d3-4cf8-972a-8e8b795eb9b1-etcd-client\") pod \"etcd-operator-b45778765-78nhb\" (UID: \"bb3d374a-97d3-4cf8-972a-8e8b795eb9b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-78nhb" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.967860 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxb62\" (UniqueName: \"kubernetes.io/projected/4674583b-142e-4968-9824-7090cfebca75-kube-api-access-jxb62\") pod \"cluster-image-registry-operator-dc59b4c8b-lzb98\" (UID: \"4674583b-142e-4968-9824-7090cfebca75\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzb98" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.967879 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/231076de-b56c-41ad-830b-8c6699d224ad-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-fq7x6\" (UID: \"231076de-b56c-41ad-830b-8c6699d224ad\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fq7x6" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.967895 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4z8q\" (UniqueName: \"kubernetes.io/projected/60b3916a-79fc-42b6-86c6-4cae051457ca-kube-api-access-r4z8q\") pod \"marketplace-operator-79b997595-289lx\" (UID: \"60b3916a-79fc-42b6-86c6-4cae051457ca\") " pod="openshift-marketplace/marketplace-operator-79b997595-289lx" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.967911 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c5ca581c-83f3-472a-962e-2edaf6e3eba9-webhook-cert\") pod \"packageserver-d55dfcdfc-jqrv2\" (UID: \"c5ca581c-83f3-472a-962e-2edaf6e3eba9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jqrv2" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.967929 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcw6m\" (UniqueName: \"kubernetes.io/projected/c522b818-1d20-4b66-afd3-098ebec69e86-kube-api-access-zcw6m\") pod \"machine-config-server-kz9wk\" (UID: \"c522b818-1d20-4b66-afd3-098ebec69e86\") " pod="openshift-machine-config-operator/machine-config-server-kz9wk" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.967943 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/bb3d374a-97d3-4cf8-972a-8e8b795eb9b1-etcd-ca\") pod \"etcd-operator-b45778765-78nhb\" (UID: \"bb3d374a-97d3-4cf8-972a-8e8b795eb9b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-78nhb" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.967958 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e457a6a-af4c-4d2b-a2af-c758de365a40-config-volume\") pod \"dns-default-tm2zm\" (UID: \"1e457a6a-af4c-4d2b-a2af-c758de365a40\") " pod="openshift-dns/dns-default-tm2zm" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.967973 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkkbd\" (UniqueName: \"kubernetes.io/projected/1e457a6a-af4c-4d2b-a2af-c758de365a40-kube-api-access-gkkbd\") pod \"dns-default-tm2zm\" (UID: \"1e457a6a-af4c-4d2b-a2af-c758de365a40\") " pod="openshift-dns/dns-default-tm2zm" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.967989 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llrdn\" (UniqueName: \"kubernetes.io/projected/279e4882-ef42-445a-891f-65dff6bdd8dc-kube-api-access-llrdn\") pod \"ingress-canary-dld8t\" (UID: \"279e4882-ef42-445a-891f-65dff6bdd8dc\") " pod="openshift-ingress-canary/ingress-canary-dld8t" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.968003 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/60b3916a-79fc-42b6-86c6-4cae051457ca-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-289lx\" (UID: \"60b3916a-79fc-42b6-86c6-4cae051457ca\") " pod="openshift-marketplace/marketplace-operator-79b997595-289lx" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.968018 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8ddf8368-8232-4211-bfca-642b1acef6a4-metrics-certs\") pod \"router-default-5444994796-ncvs7\" (UID: \"8ddf8368-8232-4211-bfca-642b1acef6a4\") " pod="openshift-ingress/router-default-5444994796-ncvs7" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.968031 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/bb3d374a-97d3-4cf8-972a-8e8b795eb9b1-etcd-service-ca\") pod \"etcd-operator-b45778765-78nhb\" (UID: \"bb3d374a-97d3-4cf8-972a-8e8b795eb9b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-78nhb" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.968045 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64703472-c8cf-4b3d-b464-a2ef7eb89c64-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-wxx9w\" (UID: \"64703472-c8cf-4b3d-b464-a2ef7eb89c64\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxx9w" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.968060 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw2zv\" (UniqueName: \"kubernetes.io/projected/8b0e59fb-3d78-4720-bcc2-2c2eb985b7bf-kube-api-access-kw2zv\") pod \"service-ca-operator-777779d784-nvs5q\" (UID: \"8b0e59fb-3d78-4720-bcc2-2c2eb985b7bf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nvs5q" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.968084 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/56d1f612-b545-4fe1-97bd-727b3d8bf062-srv-cert\") pod \"olm-operator-6b444d44fb-tw7m2\" (UID: \"56d1f612-b545-4fe1-97bd-727b3d8bf062\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tw7m2" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.968100 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4674583b-142e-4968-9824-7090cfebca75-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lzb98\" (UID: \"4674583b-142e-4968-9824-7090cfebca75\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzb98" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.968131 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c522b818-1d20-4b66-afd3-098ebec69e86-node-bootstrap-token\") pod \"machine-config-server-kz9wk\" (UID: \"c522b818-1d20-4b66-afd3-098ebec69e86\") " pod="openshift-machine-config-operator/machine-config-server-kz9wk" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.968146 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8810f12c-6dbc-4bf9-b27e-29ebc5986955-config-volume\") pod \"collect-profiles-29417265-hcswl\" (UID: \"8810f12c-6dbc-4bf9-b27e-29ebc5986955\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417265-hcswl" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.968161 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d2857e44-0b98-4f13-baca-3330474d8b44-bound-sa-token\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.968177 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e16c161d-a64d-4bc9-a223-aaf7bfc53083-proxy-tls\") pod \"machine-config-controller-84d6567774-rkmnv\" (UID: \"e16c161d-a64d-4bc9-a223-aaf7bfc53083\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rkmnv" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.968191 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb28f2df-5922-47fc-aaf0-5887ca9eb03c-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9x54c\" (UID: \"bb28f2df-5922-47fc-aaf0-5887ca9eb03c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x54c" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.968208 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnmrp\" (UniqueName: \"kubernetes.io/projected/3b977b1d-5f16-4989-b851-145fe091af2e-kube-api-access-gnmrp\") pod \"dns-operator-744455d44c-mwqcj\" (UID: \"3b977b1d-5f16-4989-b851-145fe091af2e\") " pod="openshift-dns-operator/dns-operator-744455d44c-mwqcj" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.968222 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c522b818-1d20-4b66-afd3-098ebec69e86-certs\") pod \"machine-config-server-kz9wk\" (UID: \"c522b818-1d20-4b66-afd3-098ebec69e86\") " pod="openshift-machine-config-operator/machine-config-server-kz9wk" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.968238 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/0904de59-781e-4c00-86bd-ccebf7a06587-plugins-dir\") pod \"csi-hostpathplugin-wj67k\" (UID: \"0904de59-781e-4c00-86bd-ccebf7a06587\") " pod="hostpath-provisioner/csi-hostpathplugin-wj67k" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.968254 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/935ac385-32ee-46c5-bffb-69b89b4af6a9-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-mllt2\" (UID: \"935ac385-32ee-46c5-bffb-69b89b4af6a9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mllt2" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.968311 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1883cf65-656c-439d-980d-597c40899cc5-proxy-tls\") pod \"machine-config-operator-74547568cd-2twln\" (UID: \"1883cf65-656c-439d-980d-597c40899cc5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2twln" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.968325 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e8aef227-c82b-4e7d-bf13-d99c935ad12b-srv-cert\") pod \"catalog-operator-68c6474976-bzqcl\" (UID: \"e8aef227-c82b-4e7d-bf13-d99c935ad12b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bzqcl" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.968346 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tv9q5\" (UniqueName: \"kubernetes.io/projected/36af9565-0192-4fee-9729-2bb06631e150-kube-api-access-tv9q5\") pod \"kube-storage-version-migrator-operator-b67b599dd-cbbdg\" (UID: \"36af9565-0192-4fee-9729-2bb06631e150\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cbbdg" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.968360 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/0904de59-781e-4c00-86bd-ccebf7a06587-csi-data-dir\") pod \"csi-hostpathplugin-wj67k\" (UID: \"0904de59-781e-4c00-86bd-ccebf7a06587\") " pod="hostpath-provisioner/csi-hostpathplugin-wj67k" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.968375 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36af9565-0192-4fee-9729-2bb06631e150-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-cbbdg\" (UID: \"36af9565-0192-4fee-9729-2bb06631e150\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cbbdg" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.968401 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c5ca581c-83f3-472a-962e-2edaf6e3eba9-tmpfs\") pod \"packageserver-d55dfcdfc-jqrv2\" (UID: \"c5ca581c-83f3-472a-962e-2edaf6e3eba9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jqrv2" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.968419 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7cww\" (UniqueName: \"kubernetes.io/projected/e16c161d-a64d-4bc9-a223-aaf7bfc53083-kube-api-access-p7cww\") pod \"machine-config-controller-84d6567774-rkmnv\" (UID: \"e16c161d-a64d-4bc9-a223-aaf7bfc53083\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rkmnv" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.968434 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a238598-92cd-4641-ae82-6e471740ee7f-config\") pod \"kube-apiserver-operator-766d6c64bb-85tkp\" (UID: \"9a238598-92cd-4641-ae82-6e471740ee7f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-85tkp" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.968457 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8x7k9\" (UniqueName: \"kubernetes.io/projected/8a7d392b-a4af-4e95-b50e-3e45d8c01037-kube-api-access-8x7k9\") pod \"openshift-controller-manager-operator-756b6f6bc6-6nzc7\" (UID: \"8a7d392b-a4af-4e95-b50e-3e45d8c01037\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nzc7" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.968473 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n976\" (UniqueName: \"kubernetes.io/projected/8810f12c-6dbc-4bf9-b27e-29ebc5986955-kube-api-access-6n976\") pod \"collect-profiles-29417265-hcswl\" (UID: \"8810f12c-6dbc-4bf9-b27e-29ebc5986955\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417265-hcswl" Dec 06 15:48:20 crc kubenswrapper[4813]: E1206 15:48:20.968499 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:21.468474503 +0000 UTC m=+141.359354079 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.968537 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8810f12c-6dbc-4bf9-b27e-29ebc5986955-secret-volume\") pod \"collect-profiles-29417265-hcswl\" (UID: \"8810f12c-6dbc-4bf9-b27e-29ebc5986955\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417265-hcswl" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.970569 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c5ca581c-83f3-472a-962e-2edaf6e3eba9-tmpfs\") pod \"packageserver-d55dfcdfc-jqrv2\" (UID: \"c5ca581c-83f3-472a-962e-2edaf6e3eba9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jqrv2" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.970838 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d2857e44-0b98-4f13-baca-3330474d8b44-ca-trust-extracted\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.971529 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1883cf65-656c-439d-980d-597c40899cc5-images\") pod \"machine-config-operator-74547568cd-2twln\" (UID: \"1883cf65-656c-439d-980d-597c40899cc5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2twln" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.971677 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hppjg\" (UniqueName: \"kubernetes.io/projected/231076de-b56c-41ad-830b-8c6699d224ad-kube-api-access-hppjg\") pod \"package-server-manager-789f6589d5-fq7x6\" (UID: \"231076de-b56c-41ad-830b-8c6699d224ad\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fq7x6" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.971726 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/45d5fd66-73f2-4eaa-88ad-4cc6bc597387-signing-key\") pod \"service-ca-9c57cc56f-wfhsc\" (UID: \"45d5fd66-73f2-4eaa-88ad-4cc6bc597387\") " pod="openshift-service-ca/service-ca-9c57cc56f-wfhsc" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.971756 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/0904de59-781e-4c00-86bd-ccebf7a06587-socket-dir\") pod \"csi-hostpathplugin-wj67k\" (UID: \"0904de59-781e-4c00-86bd-ccebf7a06587\") " pod="hostpath-provisioner/csi-hostpathplugin-wj67k" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.971791 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv68h\" (UniqueName: \"kubernetes.io/projected/8ddf8368-8232-4211-bfca-642b1acef6a4-kube-api-access-hv68h\") pod \"router-default-5444994796-ncvs7\" (UID: \"8ddf8368-8232-4211-bfca-642b1acef6a4\") " pod="openshift-ingress/router-default-5444994796-ncvs7" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.971808 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxcjd\" (UniqueName: \"kubernetes.io/projected/56d1f612-b545-4fe1-97bd-727b3d8bf062-kube-api-access-rxcjd\") pod \"olm-operator-6b444d44fb-tw7m2\" (UID: \"56d1f612-b545-4fe1-97bd-727b3d8bf062\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tw7m2" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.971823 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0664e3bf-bbbe-4547-9697-785efea4dd99-bound-sa-token\") pod \"ingress-operator-5b745b69d9-5cbx6\" (UID: \"0664e3bf-bbbe-4547-9697-785efea4dd99\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5cbx6" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.971842 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9549da48-ad83-4eff-878d-c9df63e66cd7-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-d4h8p\" (UID: \"9549da48-ad83-4eff-878d-c9df63e66cd7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d4h8p" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.971857 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/0904de59-781e-4c00-86bd-ccebf7a06587-registration-dir\") pod \"csi-hostpathplugin-wj67k\" (UID: \"0904de59-781e-4c00-86bd-ccebf7a06587\") " pod="hostpath-provisioner/csi-hostpathplugin-wj67k" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.971874 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d2857e44-0b98-4f13-baca-3330474d8b44-trusted-ca\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.971891 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67mng\" (UniqueName: \"kubernetes.io/projected/d2857e44-0b98-4f13-baca-3330474d8b44-kube-api-access-67mng\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.971944 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d2857e44-0b98-4f13-baca-3330474d8b44-registry-certificates\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.971961 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e16c161d-a64d-4bc9-a223-aaf7bfc53083-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-rkmnv\" (UID: \"e16c161d-a64d-4bc9-a223-aaf7bfc53083\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rkmnv" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.971980 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgr8b\" (UniqueName: \"kubernetes.io/projected/bb3d374a-97d3-4cf8-972a-8e8b795eb9b1-kube-api-access-dgr8b\") pod \"etcd-operator-b45778765-78nhb\" (UID: \"bb3d374a-97d3-4cf8-972a-8e8b795eb9b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-78nhb" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.971996 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9a238598-92cd-4641-ae82-6e471740ee7f-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-85tkp\" (UID: \"9a238598-92cd-4641-ae82-6e471740ee7f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-85tkp" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.972015 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d2857e44-0b98-4f13-baca-3330474d8b44-installation-pull-secrets\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.972030 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c5ca581c-83f3-472a-962e-2edaf6e3eba9-apiservice-cert\") pod \"packageserver-d55dfcdfc-jqrv2\" (UID: \"c5ca581c-83f3-472a-962e-2edaf6e3eba9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jqrv2" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.972045 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8ddf8368-8232-4211-bfca-642b1acef6a4-stats-auth\") pod \"router-default-5444994796-ncvs7\" (UID: \"8ddf8368-8232-4211-bfca-642b1acef6a4\") " pod="openshift-ingress/router-default-5444994796-ncvs7" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.972060 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/60b3916a-79fc-42b6-86c6-4cae051457ca-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-289lx\" (UID: \"60b3916a-79fc-42b6-86c6-4cae051457ca\") " pod="openshift-marketplace/marketplace-operator-79b997595-289lx" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.972080 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/56d1f612-b545-4fe1-97bd-727b3d8bf062-profile-collector-cert\") pod \"olm-operator-6b444d44fb-tw7m2\" (UID: \"56d1f612-b545-4fe1-97bd-727b3d8bf062\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tw7m2" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.972374 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsw2p\" (UniqueName: \"kubernetes.io/projected/0664e3bf-bbbe-4547-9697-785efea4dd99-kube-api-access-fsw2p\") pod \"ingress-operator-5b745b69d9-5cbx6\" (UID: \"0664e3bf-bbbe-4547-9697-785efea4dd99\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5cbx6" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.972588 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5vcc\" (UniqueName: \"kubernetes.io/projected/e8aef227-c82b-4e7d-bf13-d99c935ad12b-kube-api-access-g5vcc\") pod \"catalog-operator-68c6474976-bzqcl\" (UID: \"e8aef227-c82b-4e7d-bf13-d99c935ad12b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bzqcl" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.972607 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rhp6\" (UniqueName: \"kubernetes.io/projected/1883cf65-656c-439d-980d-597c40899cc5-kube-api-access-7rhp6\") pod \"machine-config-operator-74547568cd-2twln\" (UID: \"1883cf65-656c-439d-980d-597c40899cc5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2twln" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.972624 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q48dm\" (UniqueName: \"kubernetes.io/projected/0904de59-781e-4c00-86bd-ccebf7a06587-kube-api-access-q48dm\") pod \"csi-hostpathplugin-wj67k\" (UID: \"0904de59-781e-4c00-86bd-ccebf7a06587\") " pod="hostpath-provisioner/csi-hostpathplugin-wj67k" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.972640 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64703472-c8cf-4b3d-b464-a2ef7eb89c64-config\") pod \"kube-controller-manager-operator-78b949d7b-wxx9w\" (UID: \"64703472-c8cf-4b3d-b464-a2ef7eb89c64\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxx9w" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.972656 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b0e59fb-3d78-4720-bcc2-2c2eb985b7bf-serving-cert\") pod \"service-ca-operator-777779d784-nvs5q\" (UID: \"8b0e59fb-3d78-4720-bcc2-2c2eb985b7bf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nvs5q" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.973480 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d2857e44-0b98-4f13-baca-3330474d8b44-trusted-ca\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.974191 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d2857e44-0b98-4f13-baca-3330474d8b44-registry-certificates\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.972408 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4674583b-142e-4968-9824-7090cfebca75-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lzb98\" (UID: \"4674583b-142e-4968-9824-7090cfebca75\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzb98" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.978609 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4674583b-142e-4968-9824-7090cfebca75-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lzb98\" (UID: \"4674583b-142e-4968-9824-7090cfebca75\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzb98" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.978639 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdxtw\" (UniqueName: \"kubernetes.io/projected/f2dde1d1-6ebd-4592-a0cf-9a61c8b7b4ae-kube-api-access-sdxtw\") pod \"migrator-59844c95c7-c7cx6\" (UID: \"f2dde1d1-6ebd-4592-a0cf-9a61c8b7b4ae\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c7cx6" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.978708 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a7d392b-a4af-4e95-b50e-3e45d8c01037-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-6nzc7\" (UID: \"8a7d392b-a4af-4e95-b50e-3e45d8c01037\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nzc7" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.979654 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7tbr\" (UniqueName: \"kubernetes.io/projected/c5ca581c-83f3-472a-962e-2edaf6e3eba9-kube-api-access-p7tbr\") pod \"packageserver-d55dfcdfc-jqrv2\" (UID: \"c5ca581c-83f3-472a-962e-2edaf6e3eba9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jqrv2" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.979688 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/64703472-c8cf-4b3d-b464-a2ef7eb89c64-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-wxx9w\" (UID: \"64703472-c8cf-4b3d-b464-a2ef7eb89c64\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxx9w" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.979726 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0664e3bf-bbbe-4547-9697-785efea4dd99-metrics-tls\") pod \"ingress-operator-5b745b69d9-5cbx6\" (UID: \"0664e3bf-bbbe-4547-9697-785efea4dd99\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5cbx6" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.979744 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb3d374a-97d3-4cf8-972a-8e8b795eb9b1-config\") pod \"etcd-operator-b45778765-78nhb\" (UID: \"bb3d374a-97d3-4cf8-972a-8e8b795eb9b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-78nhb" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.979853 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjpjn\" (UniqueName: \"kubernetes.io/projected/26bc2fa2-9eee-4f7b-b31b-86f364df2b06-kube-api-access-zjpjn\") pod \"downloads-7954f5f757-kkbqh\" (UID: \"26bc2fa2-9eee-4f7b-b31b-86f364df2b06\") " pod="openshift-console/downloads-7954f5f757-kkbqh" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.979874 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb28f2df-5922-47fc-aaf0-5887ca9eb03c-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9x54c\" (UID: \"bb28f2df-5922-47fc-aaf0-5887ca9eb03c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x54c" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.979941 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1883cf65-656c-439d-980d-597c40899cc5-auth-proxy-config\") pod \"machine-config-operator-74547568cd-2twln\" (UID: \"1883cf65-656c-439d-980d-597c40899cc5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2twln" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.979978 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56xw6\" (UniqueName: \"kubernetes.io/projected/45d5fd66-73f2-4eaa-88ad-4cc6bc597387-kube-api-access-56xw6\") pod \"service-ca-9c57cc56f-wfhsc\" (UID: \"45d5fd66-73f2-4eaa-88ad-4cc6bc597387\") " pod="openshift-service-ca/service-ca-9c57cc56f-wfhsc" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.980466 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1883cf65-656c-439d-980d-597c40899cc5-auth-proxy-config\") pod \"machine-config-operator-74547568cd-2twln\" (UID: \"1883cf65-656c-439d-980d-597c40899cc5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2twln" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.980501 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a7d392b-a4af-4e95-b50e-3e45d8c01037-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-6nzc7\" (UID: \"8a7d392b-a4af-4e95-b50e-3e45d8c01037\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nzc7" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.980520 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/0904de59-781e-4c00-86bd-ccebf7a06587-mountpoint-dir\") pod \"csi-hostpathplugin-wj67k\" (UID: \"0904de59-781e-4c00-86bd-ccebf7a06587\") " pod="hostpath-provisioner/csi-hostpathplugin-wj67k" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.980556 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3b977b1d-5f16-4989-b851-145fe091af2e-metrics-tls\") pod \"dns-operator-744455d44c-mwqcj\" (UID: \"3b977b1d-5f16-4989-b851-145fe091af2e\") " pod="openshift-dns-operator/dns-operator-744455d44c-mwqcj" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.980574 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d2857e44-0b98-4f13-baca-3330474d8b44-registry-tls\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.981455 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a7d392b-a4af-4e95-b50e-3e45d8c01037-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-6nzc7\" (UID: \"8a7d392b-a4af-4e95-b50e-3e45d8c01037\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nzc7" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.983200 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/4674583b-142e-4968-9824-7090cfebca75-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lzb98\" (UID: \"4674583b-142e-4968-9824-7090cfebca75\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzb98" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.984519 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c5ca581c-83f3-472a-962e-2edaf6e3eba9-webhook-cert\") pod \"packageserver-d55dfcdfc-jqrv2\" (UID: \"c5ca581c-83f3-472a-962e-2edaf6e3eba9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jqrv2" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.985980 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1883cf65-656c-439d-980d-597c40899cc5-proxy-tls\") pod \"machine-config-operator-74547568cd-2twln\" (UID: \"1883cf65-656c-439d-980d-597c40899cc5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2twln" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.990977 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/4674583b-142e-4968-9824-7090cfebca75-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lzb98\" (UID: \"4674583b-142e-4968-9824-7090cfebca75\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzb98" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.991071 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a7d392b-a4af-4e95-b50e-3e45d8c01037-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-6nzc7\" (UID: \"8a7d392b-a4af-4e95-b50e-3e45d8c01037\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nzc7" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.991907 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c5ca581c-83f3-472a-962e-2edaf6e3eba9-apiservice-cert\") pod \"packageserver-d55dfcdfc-jqrv2\" (UID: \"c5ca581c-83f3-472a-962e-2edaf6e3eba9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jqrv2" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.993293 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.995013 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d2857e44-0b98-4f13-baca-3330474d8b44-registry-tls\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:20 crc kubenswrapper[4813]: I1206 15:48:20.996997 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9549da48-ad83-4eff-878d-c9df63e66cd7-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-d4h8p\" (UID: \"9549da48-ad83-4eff-878d-c9df63e66cd7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d4h8p" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.003993 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d2857e44-0b98-4f13-baca-3330474d8b44-installation-pull-secrets\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.012434 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2bzg\" (UniqueName: \"kubernetes.io/projected/9549da48-ad83-4eff-878d-c9df63e66cd7-kube-api-access-n2bzg\") pod \"cluster-samples-operator-665b6dd947-d4h8p\" (UID: \"9549da48-ad83-4eff-878d-c9df63e66cd7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d4h8p" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.015367 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8x7k9\" (UniqueName: \"kubernetes.io/projected/8a7d392b-a4af-4e95-b50e-3e45d8c01037-kube-api-access-8x7k9\") pod \"openshift-controller-manager-operator-756b6f6bc6-6nzc7\" (UID: \"8a7d392b-a4af-4e95-b50e-3e45d8c01037\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nzc7" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.015413 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d2857e44-0b98-4f13-baca-3330474d8b44-bound-sa-token\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.032056 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pmr8t"] Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.036594 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxb62\" (UniqueName: \"kubernetes.io/projected/4674583b-142e-4968-9824-7090cfebca75-kube-api-access-jxb62\") pod \"cluster-image-registry-operator-dc59b4c8b-lzb98\" (UID: \"4674583b-142e-4968-9824-7090cfebca75\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzb98" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.044790 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rhp6\" (UniqueName: \"kubernetes.io/projected/1883cf65-656c-439d-980d-597c40899cc5-kube-api-access-7rhp6\") pod \"machine-config-operator-74547568cd-2twln\" (UID: \"1883cf65-656c-439d-980d-597c40899cc5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2twln" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.045608 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8vzrm"] Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.054831 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67mng\" (UniqueName: \"kubernetes.io/projected/d2857e44-0b98-4f13-baca-3330474d8b44-kube-api-access-67mng\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.065139 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-dq6hp"] Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.092762 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4674583b-142e-4968-9824-7090cfebca75-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lzb98\" (UID: \"4674583b-142e-4968-9824-7090cfebca75\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzb98" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.092836 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c522b818-1d20-4b66-afd3-098ebec69e86-certs\") pod \"machine-config-server-kz9wk\" (UID: \"c522b818-1d20-4b66-afd3-098ebec69e86\") " pod="openshift-machine-config-operator/machine-config-server-kz9wk" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.092858 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e16c161d-a64d-4bc9-a223-aaf7bfc53083-proxy-tls\") pod \"machine-config-controller-84d6567774-rkmnv\" (UID: \"e16c161d-a64d-4bc9-a223-aaf7bfc53083\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rkmnv" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.092874 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb28f2df-5922-47fc-aaf0-5887ca9eb03c-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9x54c\" (UID: \"bb28f2df-5922-47fc-aaf0-5887ca9eb03c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x54c" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.092891 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnmrp\" (UniqueName: \"kubernetes.io/projected/3b977b1d-5f16-4989-b851-145fe091af2e-kube-api-access-gnmrp\") pod \"dns-operator-744455d44c-mwqcj\" (UID: \"3b977b1d-5f16-4989-b851-145fe091af2e\") " pod="openshift-dns-operator/dns-operator-744455d44c-mwqcj" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.092909 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/935ac385-32ee-46c5-bffb-69b89b4af6a9-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-mllt2\" (UID: \"935ac385-32ee-46c5-bffb-69b89b4af6a9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mllt2" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.092928 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/0904de59-781e-4c00-86bd-ccebf7a06587-plugins-dir\") pod \"csi-hostpathplugin-wj67k\" (UID: \"0904de59-781e-4c00-86bd-ccebf7a06587\") " pod="hostpath-provisioner/csi-hostpathplugin-wj67k" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.092952 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e8aef227-c82b-4e7d-bf13-d99c935ad12b-srv-cert\") pod \"catalog-operator-68c6474976-bzqcl\" (UID: \"e8aef227-c82b-4e7d-bf13-d99c935ad12b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bzqcl" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.092985 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093002 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tv9q5\" (UniqueName: \"kubernetes.io/projected/36af9565-0192-4fee-9729-2bb06631e150-kube-api-access-tv9q5\") pod \"kube-storage-version-migrator-operator-b67b599dd-cbbdg\" (UID: \"36af9565-0192-4fee-9729-2bb06631e150\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cbbdg" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093017 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/0904de59-781e-4c00-86bd-ccebf7a06587-csi-data-dir\") pod \"csi-hostpathplugin-wj67k\" (UID: \"0904de59-781e-4c00-86bd-ccebf7a06587\") " pod="hostpath-provisioner/csi-hostpathplugin-wj67k" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093035 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36af9565-0192-4fee-9729-2bb06631e150-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-cbbdg\" (UID: \"36af9565-0192-4fee-9729-2bb06631e150\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cbbdg" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093056 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7cww\" (UniqueName: \"kubernetes.io/projected/e16c161d-a64d-4bc9-a223-aaf7bfc53083-kube-api-access-p7cww\") pod \"machine-config-controller-84d6567774-rkmnv\" (UID: \"e16c161d-a64d-4bc9-a223-aaf7bfc53083\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rkmnv" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093072 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a238598-92cd-4641-ae82-6e471740ee7f-config\") pod \"kube-apiserver-operator-766d6c64bb-85tkp\" (UID: \"9a238598-92cd-4641-ae82-6e471740ee7f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-85tkp" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093097 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6n976\" (UniqueName: \"kubernetes.io/projected/8810f12c-6dbc-4bf9-b27e-29ebc5986955-kube-api-access-6n976\") pod \"collect-profiles-29417265-hcswl\" (UID: \"8810f12c-6dbc-4bf9-b27e-29ebc5986955\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417265-hcswl" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093111 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8810f12c-6dbc-4bf9-b27e-29ebc5986955-secret-volume\") pod \"collect-profiles-29417265-hcswl\" (UID: \"8810f12c-6dbc-4bf9-b27e-29ebc5986955\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417265-hcswl" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093124 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hppjg\" (UniqueName: \"kubernetes.io/projected/231076de-b56c-41ad-830b-8c6699d224ad-kube-api-access-hppjg\") pod \"package-server-manager-789f6589d5-fq7x6\" (UID: \"231076de-b56c-41ad-830b-8c6699d224ad\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fq7x6" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093141 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/45d5fd66-73f2-4eaa-88ad-4cc6bc597387-signing-key\") pod \"service-ca-9c57cc56f-wfhsc\" (UID: \"45d5fd66-73f2-4eaa-88ad-4cc6bc597387\") " pod="openshift-service-ca/service-ca-9c57cc56f-wfhsc" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093157 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxcjd\" (UniqueName: \"kubernetes.io/projected/56d1f612-b545-4fe1-97bd-727b3d8bf062-kube-api-access-rxcjd\") pod \"olm-operator-6b444d44fb-tw7m2\" (UID: \"56d1f612-b545-4fe1-97bd-727b3d8bf062\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tw7m2" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093170 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0664e3bf-bbbe-4547-9697-785efea4dd99-bound-sa-token\") pod \"ingress-operator-5b745b69d9-5cbx6\" (UID: \"0664e3bf-bbbe-4547-9697-785efea4dd99\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5cbx6" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093187 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/0904de59-781e-4c00-86bd-ccebf7a06587-socket-dir\") pod \"csi-hostpathplugin-wj67k\" (UID: \"0904de59-781e-4c00-86bd-ccebf7a06587\") " pod="hostpath-provisioner/csi-hostpathplugin-wj67k" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093204 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv68h\" (UniqueName: \"kubernetes.io/projected/8ddf8368-8232-4211-bfca-642b1acef6a4-kube-api-access-hv68h\") pod \"router-default-5444994796-ncvs7\" (UID: \"8ddf8368-8232-4211-bfca-642b1acef6a4\") " pod="openshift-ingress/router-default-5444994796-ncvs7" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093219 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/0904de59-781e-4c00-86bd-ccebf7a06587-registration-dir\") pod \"csi-hostpathplugin-wj67k\" (UID: \"0904de59-781e-4c00-86bd-ccebf7a06587\") " pod="hostpath-provisioner/csi-hostpathplugin-wj67k" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093235 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9a238598-92cd-4641-ae82-6e471740ee7f-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-85tkp\" (UID: \"9a238598-92cd-4641-ae82-6e471740ee7f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-85tkp" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093250 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e16c161d-a64d-4bc9-a223-aaf7bfc53083-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-rkmnv\" (UID: \"e16c161d-a64d-4bc9-a223-aaf7bfc53083\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rkmnv" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093282 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgr8b\" (UniqueName: \"kubernetes.io/projected/bb3d374a-97d3-4cf8-972a-8e8b795eb9b1-kube-api-access-dgr8b\") pod \"etcd-operator-b45778765-78nhb\" (UID: \"bb3d374a-97d3-4cf8-972a-8e8b795eb9b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-78nhb" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093297 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8ddf8368-8232-4211-bfca-642b1acef6a4-stats-auth\") pod \"router-default-5444994796-ncvs7\" (UID: \"8ddf8368-8232-4211-bfca-642b1acef6a4\") " pod="openshift-ingress/router-default-5444994796-ncvs7" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093322 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/60b3916a-79fc-42b6-86c6-4cae051457ca-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-289lx\" (UID: \"60b3916a-79fc-42b6-86c6-4cae051457ca\") " pod="openshift-marketplace/marketplace-operator-79b997595-289lx" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093345 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/56d1f612-b545-4fe1-97bd-727b3d8bf062-profile-collector-cert\") pod \"olm-operator-6b444d44fb-tw7m2\" (UID: \"56d1f612-b545-4fe1-97bd-727b3d8bf062\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tw7m2" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093363 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsw2p\" (UniqueName: \"kubernetes.io/projected/0664e3bf-bbbe-4547-9697-785efea4dd99-kube-api-access-fsw2p\") pod \"ingress-operator-5b745b69d9-5cbx6\" (UID: \"0664e3bf-bbbe-4547-9697-785efea4dd99\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5cbx6" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093381 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5vcc\" (UniqueName: \"kubernetes.io/projected/e8aef227-c82b-4e7d-bf13-d99c935ad12b-kube-api-access-g5vcc\") pod \"catalog-operator-68c6474976-bzqcl\" (UID: \"e8aef227-c82b-4e7d-bf13-d99c935ad12b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bzqcl" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093395 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q48dm\" (UniqueName: \"kubernetes.io/projected/0904de59-781e-4c00-86bd-ccebf7a06587-kube-api-access-q48dm\") pod \"csi-hostpathplugin-wj67k\" (UID: \"0904de59-781e-4c00-86bd-ccebf7a06587\") " pod="hostpath-provisioner/csi-hostpathplugin-wj67k" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093410 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64703472-c8cf-4b3d-b464-a2ef7eb89c64-config\") pod \"kube-controller-manager-operator-78b949d7b-wxx9w\" (UID: \"64703472-c8cf-4b3d-b464-a2ef7eb89c64\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxx9w" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093424 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b0e59fb-3d78-4720-bcc2-2c2eb985b7bf-serving-cert\") pod \"service-ca-operator-777779d784-nvs5q\" (UID: \"8b0e59fb-3d78-4720-bcc2-2c2eb985b7bf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nvs5q" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093440 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdxtw\" (UniqueName: \"kubernetes.io/projected/f2dde1d1-6ebd-4592-a0cf-9a61c8b7b4ae-kube-api-access-sdxtw\") pod \"migrator-59844c95c7-c7cx6\" (UID: \"f2dde1d1-6ebd-4592-a0cf-9a61c8b7b4ae\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c7cx6" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093461 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/64703472-c8cf-4b3d-b464-a2ef7eb89c64-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-wxx9w\" (UID: \"64703472-c8cf-4b3d-b464-a2ef7eb89c64\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxx9w" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093493 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0664e3bf-bbbe-4547-9697-785efea4dd99-metrics-tls\") pod \"ingress-operator-5b745b69d9-5cbx6\" (UID: \"0664e3bf-bbbe-4547-9697-785efea4dd99\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5cbx6" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093509 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb3d374a-97d3-4cf8-972a-8e8b795eb9b1-config\") pod \"etcd-operator-b45778765-78nhb\" (UID: \"bb3d374a-97d3-4cf8-972a-8e8b795eb9b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-78nhb" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093530 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb28f2df-5922-47fc-aaf0-5887ca9eb03c-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9x54c\" (UID: \"bb28f2df-5922-47fc-aaf0-5887ca9eb03c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x54c" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093547 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56xw6\" (UniqueName: \"kubernetes.io/projected/45d5fd66-73f2-4eaa-88ad-4cc6bc597387-kube-api-access-56xw6\") pod \"service-ca-9c57cc56f-wfhsc\" (UID: \"45d5fd66-73f2-4eaa-88ad-4cc6bc597387\") " pod="openshift-service-ca/service-ca-9c57cc56f-wfhsc" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093561 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3b977b1d-5f16-4989-b851-145fe091af2e-metrics-tls\") pod \"dns-operator-744455d44c-mwqcj\" (UID: \"3b977b1d-5f16-4989-b851-145fe091af2e\") " pod="openshift-dns-operator/dns-operator-744455d44c-mwqcj" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093577 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/0904de59-781e-4c00-86bd-ccebf7a06587-mountpoint-dir\") pod \"csi-hostpathplugin-wj67k\" (UID: \"0904de59-781e-4c00-86bd-ccebf7a06587\") " pod="hostpath-provisioner/csi-hostpathplugin-wj67k" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093595 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82cp2\" (UniqueName: \"kubernetes.io/projected/2e8f4e1f-b111-4e8f-8600-6ac82ded86e9-kube-api-access-82cp2\") pod \"multus-admission-controller-857f4d67dd-t8zwg\" (UID: \"2e8f4e1f-b111-4e8f-8600-6ac82ded86e9\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t8zwg" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093611 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2e8f4e1f-b111-4e8f-8600-6ac82ded86e9-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-t8zwg\" (UID: \"2e8f4e1f-b111-4e8f-8600-6ac82ded86e9\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t8zwg" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093624 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8ddf8368-8232-4211-bfca-642b1acef6a4-default-certificate\") pod \"router-default-5444994796-ncvs7\" (UID: \"8ddf8368-8232-4211-bfca-642b1acef6a4\") " pod="openshift-ingress/router-default-5444994796-ncvs7" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093639 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0664e3bf-bbbe-4547-9697-785efea4dd99-trusted-ca\") pod \"ingress-operator-5b745b69d9-5cbx6\" (UID: \"0664e3bf-bbbe-4547-9697-785efea4dd99\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5cbx6" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093661 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1e457a6a-af4c-4d2b-a2af-c758de365a40-metrics-tls\") pod \"dns-default-tm2zm\" (UID: \"1e457a6a-af4c-4d2b-a2af-c758de365a40\") " pod="openshift-dns/dns-default-tm2zm" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093675 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/45d5fd66-73f2-4eaa-88ad-4cc6bc597387-signing-cabundle\") pod \"service-ca-9c57cc56f-wfhsc\" (UID: \"45d5fd66-73f2-4eaa-88ad-4cc6bc597387\") " pod="openshift-service-ca/service-ca-9c57cc56f-wfhsc" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093693 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb28f2df-5922-47fc-aaf0-5887ca9eb03c-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9x54c\" (UID: \"bb28f2df-5922-47fc-aaf0-5887ca9eb03c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x54c" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093706 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b0e59fb-3d78-4720-bcc2-2c2eb985b7bf-config\") pod \"service-ca-operator-777779d784-nvs5q\" (UID: \"8b0e59fb-3d78-4720-bcc2-2c2eb985b7bf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nvs5q" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093720 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36af9565-0192-4fee-9729-2bb06631e150-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-cbbdg\" (UID: \"36af9565-0192-4fee-9729-2bb06631e150\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cbbdg" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093735 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6687x\" (UniqueName: \"kubernetes.io/projected/935ac385-32ee-46c5-bffb-69b89b4af6a9-kube-api-access-6687x\") pod \"control-plane-machine-set-operator-78cbb6b69f-mllt2\" (UID: \"935ac385-32ee-46c5-bffb-69b89b4af6a9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mllt2" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093749 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/279e4882-ef42-445a-891f-65dff6bdd8dc-cert\") pod \"ingress-canary-dld8t\" (UID: \"279e4882-ef42-445a-891f-65dff6bdd8dc\") " pod="openshift-ingress-canary/ingress-canary-dld8t" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093763 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e8aef227-c82b-4e7d-bf13-d99c935ad12b-profile-collector-cert\") pod \"catalog-operator-68c6474976-bzqcl\" (UID: \"e8aef227-c82b-4e7d-bf13-d99c935ad12b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bzqcl" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093778 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb3d374a-97d3-4cf8-972a-8e8b795eb9b1-serving-cert\") pod \"etcd-operator-b45778765-78nhb\" (UID: \"bb3d374a-97d3-4cf8-972a-8e8b795eb9b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-78nhb" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093793 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ddf8368-8232-4211-bfca-642b1acef6a4-service-ca-bundle\") pod \"router-default-5444994796-ncvs7\" (UID: \"8ddf8368-8232-4211-bfca-642b1acef6a4\") " pod="openshift-ingress/router-default-5444994796-ncvs7" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093807 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a238598-92cd-4641-ae82-6e471740ee7f-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-85tkp\" (UID: \"9a238598-92cd-4641-ae82-6e471740ee7f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-85tkp" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093828 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bb3d374a-97d3-4cf8-972a-8e8b795eb9b1-etcd-client\") pod \"etcd-operator-b45778765-78nhb\" (UID: \"bb3d374a-97d3-4cf8-972a-8e8b795eb9b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-78nhb" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093844 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4z8q\" (UniqueName: \"kubernetes.io/projected/60b3916a-79fc-42b6-86c6-4cae051457ca-kube-api-access-r4z8q\") pod \"marketplace-operator-79b997595-289lx\" (UID: \"60b3916a-79fc-42b6-86c6-4cae051457ca\") " pod="openshift-marketplace/marketplace-operator-79b997595-289lx" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093860 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/231076de-b56c-41ad-830b-8c6699d224ad-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-fq7x6\" (UID: \"231076de-b56c-41ad-830b-8c6699d224ad\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fq7x6" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093875 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e457a6a-af4c-4d2b-a2af-c758de365a40-config-volume\") pod \"dns-default-tm2zm\" (UID: \"1e457a6a-af4c-4d2b-a2af-c758de365a40\") " pod="openshift-dns/dns-default-tm2zm" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093890 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkkbd\" (UniqueName: \"kubernetes.io/projected/1e457a6a-af4c-4d2b-a2af-c758de365a40-kube-api-access-gkkbd\") pod \"dns-default-tm2zm\" (UID: \"1e457a6a-af4c-4d2b-a2af-c758de365a40\") " pod="openshift-dns/dns-default-tm2zm" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093906 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcw6m\" (UniqueName: \"kubernetes.io/projected/c522b818-1d20-4b66-afd3-098ebec69e86-kube-api-access-zcw6m\") pod \"machine-config-server-kz9wk\" (UID: \"c522b818-1d20-4b66-afd3-098ebec69e86\") " pod="openshift-machine-config-operator/machine-config-server-kz9wk" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093919 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/bb3d374a-97d3-4cf8-972a-8e8b795eb9b1-etcd-ca\") pod \"etcd-operator-b45778765-78nhb\" (UID: \"bb3d374a-97d3-4cf8-972a-8e8b795eb9b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-78nhb" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093933 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8ddf8368-8232-4211-bfca-642b1acef6a4-metrics-certs\") pod \"router-default-5444994796-ncvs7\" (UID: \"8ddf8368-8232-4211-bfca-642b1acef6a4\") " pod="openshift-ingress/router-default-5444994796-ncvs7" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093947 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llrdn\" (UniqueName: \"kubernetes.io/projected/279e4882-ef42-445a-891f-65dff6bdd8dc-kube-api-access-llrdn\") pod \"ingress-canary-dld8t\" (UID: \"279e4882-ef42-445a-891f-65dff6bdd8dc\") " pod="openshift-ingress-canary/ingress-canary-dld8t" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093962 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/60b3916a-79fc-42b6-86c6-4cae051457ca-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-289lx\" (UID: \"60b3916a-79fc-42b6-86c6-4cae051457ca\") " pod="openshift-marketplace/marketplace-operator-79b997595-289lx" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093976 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/bb3d374a-97d3-4cf8-972a-8e8b795eb9b1-etcd-service-ca\") pod \"etcd-operator-b45778765-78nhb\" (UID: \"bb3d374a-97d3-4cf8-972a-8e8b795eb9b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-78nhb" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.093990 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64703472-c8cf-4b3d-b464-a2ef7eb89c64-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-wxx9w\" (UID: \"64703472-c8cf-4b3d-b464-a2ef7eb89c64\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxx9w" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.094004 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw2zv\" (UniqueName: \"kubernetes.io/projected/8b0e59fb-3d78-4720-bcc2-2c2eb985b7bf-kube-api-access-kw2zv\") pod \"service-ca-operator-777779d784-nvs5q\" (UID: \"8b0e59fb-3d78-4720-bcc2-2c2eb985b7bf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nvs5q" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.094019 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/56d1f612-b545-4fe1-97bd-727b3d8bf062-srv-cert\") pod \"olm-operator-6b444d44fb-tw7m2\" (UID: \"56d1f612-b545-4fe1-97bd-727b3d8bf062\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tw7m2" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.094033 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c522b818-1d20-4b66-afd3-098ebec69e86-node-bootstrap-token\") pod \"machine-config-server-kz9wk\" (UID: \"c522b818-1d20-4b66-afd3-098ebec69e86\") " pod="openshift-machine-config-operator/machine-config-server-kz9wk" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.094047 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8810f12c-6dbc-4bf9-b27e-29ebc5986955-config-volume\") pod \"collect-profiles-29417265-hcswl\" (UID: \"8810f12c-6dbc-4bf9-b27e-29ebc5986955\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417265-hcswl" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.094351 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p"] Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.097804 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7tbr\" (UniqueName: \"kubernetes.io/projected/c5ca581c-83f3-472a-962e-2edaf6e3eba9-kube-api-access-p7tbr\") pod \"packageserver-d55dfcdfc-jqrv2\" (UID: \"c5ca581c-83f3-472a-962e-2edaf6e3eba9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jqrv2" Dec 06 15:48:21 crc kubenswrapper[4813]: E1206 15:48:21.097866 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:21.597847507 +0000 UTC m=+141.488727083 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.098082 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/0904de59-781e-4c00-86bd-ccebf7a06587-socket-dir\") pod \"csi-hostpathplugin-wj67k\" (UID: \"0904de59-781e-4c00-86bd-ccebf7a06587\") " pod="hostpath-provisioner/csi-hostpathplugin-wj67k" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.098163 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/0904de59-781e-4c00-86bd-ccebf7a06587-registration-dir\") pod \"csi-hostpathplugin-wj67k\" (UID: \"0904de59-781e-4c00-86bd-ccebf7a06587\") " pod="hostpath-provisioner/csi-hostpathplugin-wj67k" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.098834 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e16c161d-a64d-4bc9-a223-aaf7bfc53083-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-rkmnv\" (UID: \"e16c161d-a64d-4bc9-a223-aaf7bfc53083\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rkmnv" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.098944 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8810f12c-6dbc-4bf9-b27e-29ebc5986955-config-volume\") pod \"collect-profiles-29417265-hcswl\" (UID: \"8810f12c-6dbc-4bf9-b27e-29ebc5986955\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417265-hcswl" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.101305 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/bb3d374a-97d3-4cf8-972a-8e8b795eb9b1-etcd-ca\") pod \"etcd-operator-b45778765-78nhb\" (UID: \"bb3d374a-97d3-4cf8-972a-8e8b795eb9b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-78nhb" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.108295 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c522b818-1d20-4b66-afd3-098ebec69e86-certs\") pod \"machine-config-server-kz9wk\" (UID: \"c522b818-1d20-4b66-afd3-098ebec69e86\") " pod="openshift-machine-config-operator/machine-config-server-kz9wk" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.116681 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ddf8368-8232-4211-bfca-642b1acef6a4-service-ca-bundle\") pod \"router-default-5444994796-ncvs7\" (UID: \"8ddf8368-8232-4211-bfca-642b1acef6a4\") " pod="openshift-ingress/router-default-5444994796-ncvs7" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.118925 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8ddf8368-8232-4211-bfca-642b1acef6a4-stats-auth\") pod \"router-default-5444994796-ncvs7\" (UID: \"8ddf8368-8232-4211-bfca-642b1acef6a4\") " pod="openshift-ingress/router-default-5444994796-ncvs7" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.122022 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0664e3bf-bbbe-4547-9697-785efea4dd99-trusted-ca\") pod \"ingress-operator-5b745b69d9-5cbx6\" (UID: \"0664e3bf-bbbe-4547-9697-785efea4dd99\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5cbx6" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.122930 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a238598-92cd-4641-ae82-6e471740ee7f-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-85tkp\" (UID: \"9a238598-92cd-4641-ae82-6e471740ee7f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-85tkp" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.125241 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bb3d374a-97d3-4cf8-972a-8e8b795eb9b1-etcd-client\") pod \"etcd-operator-b45778765-78nhb\" (UID: \"bb3d374a-97d3-4cf8-972a-8e8b795eb9b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-78nhb" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.128764 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/231076de-b56c-41ad-830b-8c6699d224ad-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-fq7x6\" (UID: \"231076de-b56c-41ad-830b-8c6699d224ad\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fq7x6" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.128782 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb3d374a-97d3-4cf8-972a-8e8b795eb9b1-serving-cert\") pod \"etcd-operator-b45778765-78nhb\" (UID: \"bb3d374a-97d3-4cf8-972a-8e8b795eb9b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-78nhb" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.128988 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/0904de59-781e-4c00-86bd-ccebf7a06587-mountpoint-dir\") pod \"csi-hostpathplugin-wj67k\" (UID: \"0904de59-781e-4c00-86bd-ccebf7a06587\") " pod="hostpath-provisioner/csi-hostpathplugin-wj67k" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.130228 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36af9565-0192-4fee-9729-2bb06631e150-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-cbbdg\" (UID: \"36af9565-0192-4fee-9729-2bb06631e150\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cbbdg" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.131709 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e8aef227-c82b-4e7d-bf13-d99c935ad12b-profile-collector-cert\") pod \"catalog-operator-68c6474976-bzqcl\" (UID: \"e8aef227-c82b-4e7d-bf13-d99c935ad12b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bzqcl" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.132243 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjpjn\" (UniqueName: \"kubernetes.io/projected/26bc2fa2-9eee-4f7b-b31b-86f364df2b06-kube-api-access-zjpjn\") pod \"downloads-7954f5f757-kkbqh\" (UID: \"26bc2fa2-9eee-4f7b-b31b-86f364df2b06\") " pod="openshift-console/downloads-7954f5f757-kkbqh" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.132928 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64703472-c8cf-4b3d-b464-a2ef7eb89c64-config\") pod \"kube-controller-manager-operator-78b949d7b-wxx9w\" (UID: \"64703472-c8cf-4b3d-b464-a2ef7eb89c64\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxx9w" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.137254 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/0904de59-781e-4c00-86bd-ccebf7a06587-csi-data-dir\") pod \"csi-hostpathplugin-wj67k\" (UID: \"0904de59-781e-4c00-86bd-ccebf7a06587\") " pod="hostpath-provisioner/csi-hostpathplugin-wj67k" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.140782 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0664e3bf-bbbe-4547-9697-785efea4dd99-metrics-tls\") pod \"ingress-operator-5b745b69d9-5cbx6\" (UID: \"0664e3bf-bbbe-4547-9697-785efea4dd99\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5cbx6" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.141901 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/0904de59-781e-4c00-86bd-ccebf7a06587-plugins-dir\") pod \"csi-hostpathplugin-wj67k\" (UID: \"0904de59-781e-4c00-86bd-ccebf7a06587\") " pod="hostpath-provisioner/csi-hostpathplugin-wj67k" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.143122 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d4h8p" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.143415 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36af9565-0192-4fee-9729-2bb06631e150-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-cbbdg\" (UID: \"36af9565-0192-4fee-9729-2bb06631e150\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cbbdg" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.145072 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb3d374a-97d3-4cf8-972a-8e8b795eb9b1-config\") pod \"etcd-operator-b45778765-78nhb\" (UID: \"bb3d374a-97d3-4cf8-972a-8e8b795eb9b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-78nhb" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.145250 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb28f2df-5922-47fc-aaf0-5887ca9eb03c-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9x54c\" (UID: \"bb28f2df-5922-47fc-aaf0-5887ca9eb03c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x54c" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.146391 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/60b3916a-79fc-42b6-86c6-4cae051457ca-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-289lx\" (UID: \"60b3916a-79fc-42b6-86c6-4cae051457ca\") " pod="openshift-marketplace/marketplace-operator-79b997595-289lx" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.146537 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb28f2df-5922-47fc-aaf0-5887ca9eb03c-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9x54c\" (UID: \"bb28f2df-5922-47fc-aaf0-5887ca9eb03c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x54c" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.146833 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/935ac385-32ee-46c5-bffb-69b89b4af6a9-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-mllt2\" (UID: \"935ac385-32ee-46c5-bffb-69b89b4af6a9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mllt2" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.147806 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/45d5fd66-73f2-4eaa-88ad-4cc6bc597387-signing-cabundle\") pod \"service-ca-9c57cc56f-wfhsc\" (UID: \"45d5fd66-73f2-4eaa-88ad-4cc6bc597387\") " pod="openshift-service-ca/service-ca-9c57cc56f-wfhsc" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.150165 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a238598-92cd-4641-ae82-6e471740ee7f-config\") pod \"kube-apiserver-operator-766d6c64bb-85tkp\" (UID: \"9a238598-92cd-4641-ae82-6e471740ee7f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-85tkp" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.150545 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64703472-c8cf-4b3d-b464-a2ef7eb89c64-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-wxx9w\" (UID: \"64703472-c8cf-4b3d-b464-a2ef7eb89c64\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxx9w" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.150780 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e16c161d-a64d-4bc9-a223-aaf7bfc53083-proxy-tls\") pod \"machine-config-controller-84d6567774-rkmnv\" (UID: \"e16c161d-a64d-4bc9-a223-aaf7bfc53083\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rkmnv" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.151212 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/bb3d374a-97d3-4cf8-972a-8e8b795eb9b1-etcd-service-ca\") pod \"etcd-operator-b45778765-78nhb\" (UID: \"bb3d374a-97d3-4cf8-972a-8e8b795eb9b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-78nhb" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.151867 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b0e59fb-3d78-4720-bcc2-2c2eb985b7bf-config\") pod \"service-ca-operator-777779d784-nvs5q\" (UID: \"8b0e59fb-3d78-4720-bcc2-2c2eb985b7bf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nvs5q" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.168798 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/45d5fd66-73f2-4eaa-88ad-4cc6bc597387-signing-key\") pod \"service-ca-9c57cc56f-wfhsc\" (UID: \"45d5fd66-73f2-4eaa-88ad-4cc6bc597387\") " pod="openshift-service-ca/service-ca-9c57cc56f-wfhsc" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.169296 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/56d1f612-b545-4fe1-97bd-727b3d8bf062-profile-collector-cert\") pod \"olm-operator-6b444d44fb-tw7m2\" (UID: \"56d1f612-b545-4fe1-97bd-727b3d8bf062\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tw7m2" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.169551 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e457a6a-af4c-4d2b-a2af-c758de365a40-config-volume\") pod \"dns-default-tm2zm\" (UID: \"1e457a6a-af4c-4d2b-a2af-c758de365a40\") " pod="openshift-dns/dns-default-tm2zm" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.171203 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8ddf8368-8232-4211-bfca-642b1acef6a4-default-certificate\") pod \"router-default-5444994796-ncvs7\" (UID: \"8ddf8368-8232-4211-bfca-642b1acef6a4\") " pod="openshift-ingress/router-default-5444994796-ncvs7" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.171541 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2e8f4e1f-b111-4e8f-8600-6ac82ded86e9-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-t8zwg\" (UID: \"2e8f4e1f-b111-4e8f-8600-6ac82ded86e9\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t8zwg" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.172008 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c522b818-1d20-4b66-afd3-098ebec69e86-node-bootstrap-token\") pod \"machine-config-server-kz9wk\" (UID: \"c522b818-1d20-4b66-afd3-098ebec69e86\") " pod="openshift-machine-config-operator/machine-config-server-kz9wk" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.172397 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8ddf8368-8232-4211-bfca-642b1acef6a4-metrics-certs\") pod \"router-default-5444994796-ncvs7\" (UID: \"8ddf8368-8232-4211-bfca-642b1acef6a4\") " pod="openshift-ingress/router-default-5444994796-ncvs7" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.173012 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3b977b1d-5f16-4989-b851-145fe091af2e-metrics-tls\") pod \"dns-operator-744455d44c-mwqcj\" (UID: \"3b977b1d-5f16-4989-b851-145fe091af2e\") " pod="openshift-dns-operator/dns-operator-744455d44c-mwqcj" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.173890 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/279e4882-ef42-445a-891f-65dff6bdd8dc-cert\") pod \"ingress-canary-dld8t\" (UID: \"279e4882-ef42-445a-891f-65dff6bdd8dc\") " pod="openshift-ingress-canary/ingress-canary-dld8t" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.173944 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q"] Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.176067 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b0e59fb-3d78-4720-bcc2-2c2eb985b7bf-serving-cert\") pod \"service-ca-operator-777779d784-nvs5q\" (UID: \"8b0e59fb-3d78-4720-bcc2-2c2eb985b7bf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nvs5q" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.176277 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/56d1f612-b545-4fe1-97bd-727b3d8bf062-srv-cert\") pod \"olm-operator-6b444d44fb-tw7m2\" (UID: \"56d1f612-b545-4fe1-97bd-727b3d8bf062\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tw7m2" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.176840 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e8aef227-c82b-4e7d-bf13-d99c935ad12b-srv-cert\") pod \"catalog-operator-68c6474976-bzqcl\" (UID: \"e8aef227-c82b-4e7d-bf13-d99c935ad12b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bzqcl" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.177850 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/64703472-c8cf-4b3d-b464-a2ef7eb89c64-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-wxx9w\" (UID: \"64703472-c8cf-4b3d-b464-a2ef7eb89c64\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxx9w" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.180235 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnmrp\" (UniqueName: \"kubernetes.io/projected/3b977b1d-5f16-4989-b851-145fe091af2e-kube-api-access-gnmrp\") pod \"dns-operator-744455d44c-mwqcj\" (UID: \"3b977b1d-5f16-4989-b851-145fe091af2e\") " pod="openshift-dns-operator/dns-operator-744455d44c-mwqcj" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.180421 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8810f12c-6dbc-4bf9-b27e-29ebc5986955-secret-volume\") pod \"collect-profiles-29417265-hcswl\" (UID: \"8810f12c-6dbc-4bf9-b27e-29ebc5986955\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417265-hcswl" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.180698 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1e457a6a-af4c-4d2b-a2af-c758de365a40-metrics-tls\") pod \"dns-default-tm2zm\" (UID: \"1e457a6a-af4c-4d2b-a2af-c758de365a40\") " pod="openshift-dns/dns-default-tm2zm" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.186273 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/60b3916a-79fc-42b6-86c6-4cae051457ca-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-289lx\" (UID: \"60b3916a-79fc-42b6-86c6-4cae051457ca\") " pod="openshift-marketplace/marketplace-operator-79b997595-289lx" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.192426 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzb98" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.193821 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxcjd\" (UniqueName: \"kubernetes.io/projected/56d1f612-b545-4fe1-97bd-727b3d8bf062-kube-api-access-rxcjd\") pod \"olm-operator-6b444d44fb-tw7m2\" (UID: \"56d1f612-b545-4fe1-97bd-727b3d8bf062\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tw7m2" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.194522 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:21 crc kubenswrapper[4813]: E1206 15:48:21.194761 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:21.694736842 +0000 UTC m=+141.585616418 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.194930 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:21 crc kubenswrapper[4813]: E1206 15:48:21.195350 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:21.695331867 +0000 UTC m=+141.586211443 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.207832 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nzc7" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.213983 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jqrv2" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.224555 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2twln" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.232135 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rtsv4"] Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.233034 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0664e3bf-bbbe-4547-9697-785efea4dd99-bound-sa-token\") pod \"ingress-operator-5b745b69d9-5cbx6\" (UID: \"0664e3bf-bbbe-4547-9697-785efea4dd99\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5cbx6" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.235900 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv68h\" (UniqueName: \"kubernetes.io/projected/8ddf8368-8232-4211-bfca-642b1acef6a4-kube-api-access-hv68h\") pod \"router-default-5444994796-ncvs7\" (UID: \"8ddf8368-8232-4211-bfca-642b1acef6a4\") " pod="openshift-ingress/router-default-5444994796-ncvs7" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.249038 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxx9w" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.252275 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9a238598-92cd-4641-ae82-6e471740ee7f-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-85tkp\" (UID: \"9a238598-92cd-4641-ae82-6e471740ee7f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-85tkp" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.268212 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-85tkp" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.275817 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgr8b\" (UniqueName: \"kubernetes.io/projected/bb3d374a-97d3-4cf8-972a-8e8b795eb9b1-kube-api-access-dgr8b\") pod \"etcd-operator-b45778765-78nhb\" (UID: \"bb3d374a-97d3-4cf8-972a-8e8b795eb9b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-78nhb" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.281250 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-mwqcj" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.295647 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:21 crc kubenswrapper[4813]: E1206 15:48:21.296893 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:21.796874055 +0000 UTC m=+141.687753641 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.301475 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-5t7m6"] Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.303558 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4z8q\" (UniqueName: \"kubernetes.io/projected/60b3916a-79fc-42b6-86c6-4cae051457ca-kube-api-access-r4z8q\") pod \"marketplace-operator-79b997595-289lx\" (UID: \"60b3916a-79fc-42b6-86c6-4cae051457ca\") " pod="openshift-marketplace/marketplace-operator-79b997595-289lx" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.314815 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hppjg\" (UniqueName: \"kubernetes.io/projected/231076de-b56c-41ad-830b-8c6699d224ad-kube-api-access-hppjg\") pod \"package-server-manager-789f6589d5-fq7x6\" (UID: \"231076de-b56c-41ad-830b-8c6699d224ad\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fq7x6" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.318625 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tw7m2" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.334421 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkkbd\" (UniqueName: \"kubernetes.io/projected/1e457a6a-af4c-4d2b-a2af-c758de365a40-kube-api-access-gkkbd\") pod \"dns-default-tm2zm\" (UID: \"1e457a6a-af4c-4d2b-a2af-c758de365a40\") " pod="openshift-dns/dns-default-tm2zm" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.337609 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fszk7"] Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.342682 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-ncvs7" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.351434 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcw6m\" (UniqueName: \"kubernetes.io/projected/c522b818-1d20-4b66-afd3-098ebec69e86-kube-api-access-zcw6m\") pod \"machine-config-server-kz9wk\" (UID: \"c522b818-1d20-4b66-afd3-098ebec69e86\") " pod="openshift-machine-config-operator/machine-config-server-kz9wk" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.392382 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-tm2zm" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.394810 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56xw6\" (UniqueName: \"kubernetes.io/projected/45d5fd66-73f2-4eaa-88ad-4cc6bc597387-kube-api-access-56xw6\") pod \"service-ca-9c57cc56f-wfhsc\" (UID: \"45d5fd66-73f2-4eaa-88ad-4cc6bc597387\") " pod="openshift-service-ca/service-ca-9c57cc56f-wfhsc" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.395881 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-wfhsc" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.407782 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.408649 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-kz9wk" Dec 06 15:48:21 crc kubenswrapper[4813]: E1206 15:48:21.408914 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:21.90889619 +0000 UTC m=+141.799775766 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.415875 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsw2p\" (UniqueName: \"kubernetes.io/projected/0664e3bf-bbbe-4547-9697-785efea4dd99-kube-api-access-fsw2p\") pod \"ingress-operator-5b745b69d9-5cbx6\" (UID: \"0664e3bf-bbbe-4547-9697-785efea4dd99\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5cbx6" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.434930 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-kkbqh" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.449273 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tv9q5\" (UniqueName: \"kubernetes.io/projected/36af9565-0192-4fee-9729-2bb06631e150-kube-api-access-tv9q5\") pod \"kube-storage-version-migrator-operator-b67b599dd-cbbdg\" (UID: \"36af9565-0192-4fee-9729-2bb06631e150\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cbbdg" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.473397 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5vcc\" (UniqueName: \"kubernetes.io/projected/e8aef227-c82b-4e7d-bf13-d99c935ad12b-kube-api-access-g5vcc\") pod \"catalog-operator-68c6474976-bzqcl\" (UID: \"e8aef227-c82b-4e7d-bf13-d99c935ad12b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bzqcl" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.481252 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82cp2\" (UniqueName: \"kubernetes.io/projected/2e8f4e1f-b111-4e8f-8600-6ac82ded86e9-kube-api-access-82cp2\") pod \"multus-admission-controller-857f4d67dd-t8zwg\" (UID: \"2e8f4e1f-b111-4e8f-8600-6ac82ded86e9\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t8zwg" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.481864 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-klqt2"] Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.502883 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6687x\" (UniqueName: \"kubernetes.io/projected/935ac385-32ee-46c5-bffb-69b89b4af6a9-kube-api-access-6687x\") pod \"control-plane-machine-set-operator-78cbb6b69f-mllt2\" (UID: \"935ac385-32ee-46c5-bffb-69b89b4af6a9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mllt2" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.513385 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:21 crc kubenswrapper[4813]: E1206 15:48:21.513668 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:22.013635942 +0000 UTC m=+141.904515518 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:21 crc kubenswrapper[4813]: W1206 15:48:21.515351 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd57f80ec_52ac_409d_9dc1_93a89d59f714.slice/crio-972af641dddb14451922cc1f7dfa54cfa3aaddd05bdbd4e6c0cf8fbfcb480295 WatchSource:0}: Error finding container 972af641dddb14451922cc1f7dfa54cfa3aaddd05bdbd4e6c0cf8fbfcb480295: Status 404 returned error can't find the container with id 972af641dddb14451922cc1f7dfa54cfa3aaddd05bdbd4e6c0cf8fbfcb480295 Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.516567 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q48dm\" (UniqueName: \"kubernetes.io/projected/0904de59-781e-4c00-86bd-ccebf7a06587-kube-api-access-q48dm\") pod \"csi-hostpathplugin-wj67k\" (UID: \"0904de59-781e-4c00-86bd-ccebf7a06587\") " pod="hostpath-provisioner/csi-hostpathplugin-wj67k" Dec 06 15:48:21 crc kubenswrapper[4813]: W1206 15:48:21.521580 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod994c44ca_f520_4452_b267_961bd400908e.slice/crio-016f963155111634ee3a14bfc168a30594f6053f636547169d7e28b46ff766a7 WatchSource:0}: Error finding container 016f963155111634ee3a14bfc168a30594f6053f636547169d7e28b46ff766a7: Status 404 returned error can't find the container with id 016f963155111634ee3a14bfc168a30594f6053f636547169d7e28b46ff766a7 Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.522011 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdxtw\" (UniqueName: \"kubernetes.io/projected/f2dde1d1-6ebd-4592-a0cf-9a61c8b7b4ae-kube-api-access-sdxtw\") pod \"migrator-59844c95c7-c7cx6\" (UID: \"f2dde1d1-6ebd-4592-a0cf-9a61c8b7b4ae\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c7cx6" Dec 06 15:48:21 crc kubenswrapper[4813]: W1206 15:48:21.532771 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f02af08_d7ad_4460_b103_cf44d7598449.slice/crio-a68fb4e04e6ee20a9e4c15df6fc9023a25d7a6f95d00706f1db52388d5d32dad WatchSource:0}: Error finding container a68fb4e04e6ee20a9e4c15df6fc9023a25d7a6f95d00706f1db52388d5d32dad: Status 404 returned error can't find the container with id a68fb4e04e6ee20a9e4c15df6fc9023a25d7a6f95d00706f1db52388d5d32dad Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.543079 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fq7x6" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.547914 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw2zv\" (UniqueName: \"kubernetes.io/projected/8b0e59fb-3d78-4720-bcc2-2c2eb985b7bf-kube-api-access-kw2zv\") pod \"service-ca-operator-777779d784-nvs5q\" (UID: \"8b0e59fb-3d78-4720-bcc2-2c2eb985b7bf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nvs5q" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.559694 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llrdn\" (UniqueName: \"kubernetes.io/projected/279e4882-ef42-445a-891f-65dff6bdd8dc-kube-api-access-llrdn\") pod \"ingress-canary-dld8t\" (UID: \"279e4882-ef42-445a-891f-65dff6bdd8dc\") " pod="openshift-ingress-canary/ingress-canary-dld8t" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.560214 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-78nhb" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.576496 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-t8zwg" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.592152 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-289lx" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.594675 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bzqcl" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.594752 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n976\" (UniqueName: \"kubernetes.io/projected/8810f12c-6dbc-4bf9-b27e-29ebc5986955-kube-api-access-6n976\") pod \"collect-profiles-29417265-hcswl\" (UID: \"8810f12c-6dbc-4bf9-b27e-29ebc5986955\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417265-hcswl" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.602385 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mllt2" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.632502 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cbbdg" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.632513 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nvs5q" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.633294 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:21 crc kubenswrapper[4813]: E1206 15:48:21.633574 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:22.133563117 +0000 UTC m=+142.024442693 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.655209 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5cbx6" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.662485 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7cww\" (UniqueName: \"kubernetes.io/projected/e16c161d-a64d-4bc9-a223-aaf7bfc53083-kube-api-access-p7cww\") pod \"machine-config-controller-84d6567774-rkmnv\" (UID: \"e16c161d-a64d-4bc9-a223-aaf7bfc53083\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rkmnv" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.666714 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c7cx6" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.697468 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-dld8t" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.716624 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d4h8p"] Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.739180 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:21 crc kubenswrapper[4813]: E1206 15:48:21.739598 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:22.239583083 +0000 UTC m=+142.130462659 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.740333 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-wj67k" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.891620 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417265-hcswl" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.892520 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:21 crc kubenswrapper[4813]: E1206 15:48:21.892796 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:22.392785518 +0000 UTC m=+142.283665094 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.910309 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb28f2df-5922-47fc-aaf0-5887ca9eb03c-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9x54c\" (UID: \"bb28f2df-5922-47fc-aaf0-5887ca9eb03c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x54c" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.924788 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rkmnv" Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.925618 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-5t7m6" event={"ID":"d57f80ec-52ac-409d-9dc1-93a89d59f714","Type":"ContainerStarted","Data":"972af641dddb14451922cc1f7dfa54cfa3aaddd05bdbd4e6c0cf8fbfcb480295"} Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.929160 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q" event={"ID":"4b409627-4716-4bdd-a660-4f88693150a2","Type":"ContainerStarted","Data":"2c79dfb191db5625286c5f1e6423c158411cee10845f2895cbd98f97f526f35f"} Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.930204 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jqrv2"] Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.932510 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-dq6hp" event={"ID":"0db3a3aa-f872-43c3-ade2-58c55a448fe3","Type":"ContainerStarted","Data":"0bb737dd9022d8e9ca607775b35e9f85ca6ea9fd12ec47b3229a07057d7731d2"} Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.933554 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-klqt2" event={"ID":"2f02af08-d7ad-4460-b103-cf44d7598449","Type":"ContainerStarted","Data":"a68fb4e04e6ee20a9e4c15df6fc9023a25d7a6f95d00706f1db52388d5d32dad"} Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.934543 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rtsv4" event={"ID":"8c420468-f370-44cb-9e39-7ae8f12b461e","Type":"ContainerStarted","Data":"8a06394c5e71da815b410af157a22ec387f5ee2f1aa9fb1419cdcc69a1d7d2b2"} Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.940702 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-6bbs5" event={"ID":"1d5678dc-7a74-4fff-818c-e34321037686","Type":"ContainerStarted","Data":"fd283e9e9706af8249ef143f4974d9053373dd945b4e3593bec104acbcc002c2"} Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.942946 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" event={"ID":"a1353568-e317-4a7d-8765-675821dac72b","Type":"ContainerStarted","Data":"e6b7f0668977010e574f3f8833a281c92d5bc3a17770aa93c3635651b2afc484"} Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.949810 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-gb6cc" event={"ID":"36f25c5a-3e57-4367-b306-db9661e4f7c9","Type":"ContainerStarted","Data":"4861be593d3b9a7a02282408617c25a42bd7fd31cdca7fb4fd51a31d05090c08"} Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.953279 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" event={"ID":"994c44ca-f520-4452-b267-961bd400908e","Type":"ContainerStarted","Data":"016f963155111634ee3a14bfc168a30594f6053f636547169d7e28b46ff766a7"} Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.954174 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pmr8t" event={"ID":"30e8f6c3-6c7e-45dc-9a6f-e2de13d1322c","Type":"ContainerStarted","Data":"c36834ad69de7c739900992c291f0fcb457afd146eb3b235288f7933276aa5ac"} Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.955146 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" event={"ID":"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a","Type":"ContainerStarted","Data":"78e88474a524b83bc2cd87dc463910d358e30429d2221b319c11ccc4783526a5"} Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.993776 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:21 crc kubenswrapper[4813]: E1206 15:48:21.998783 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:22.498752773 +0000 UTC m=+142.389632339 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:21 crc kubenswrapper[4813]: I1206 15:48:21.999511 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:22 crc kubenswrapper[4813]: E1206 15:48:22.000616 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:22.500598462 +0000 UTC m=+142.391478048 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:22 crc kubenswrapper[4813]: I1206 15:48:22.069817 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxx9w"] Dec 06 15:48:22 crc kubenswrapper[4813]: I1206 15:48:22.101580 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:22 crc kubenswrapper[4813]: E1206 15:48:22.101863 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:22.601848312 +0000 UTC m=+142.492727888 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:22 crc kubenswrapper[4813]: I1206 15:48:22.136897 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x54c" Dec 06 15:48:22 crc kubenswrapper[4813]: I1206 15:48:22.222510 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:22 crc kubenswrapper[4813]: E1206 15:48:22.222840 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:22.722828384 +0000 UTC m=+142.613707960 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:22 crc kubenswrapper[4813]: I1206 15:48:22.325138 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:22 crc kubenswrapper[4813]: E1206 15:48:22.325944 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:22.825910703 +0000 UTC m=+142.716790279 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:22 crc kubenswrapper[4813]: I1206 15:48:22.531436 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:22 crc kubenswrapper[4813]: E1206 15:48:22.531973 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:23.031960767 +0000 UTC m=+142.922840343 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:22 crc kubenswrapper[4813]: I1206 15:48:22.541932 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nzc7"] Dec 06 15:48:22 crc kubenswrapper[4813]: I1206 15:48:22.555549 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzb98"] Dec 06 15:48:22 crc kubenswrapper[4813]: I1206 15:48:22.632453 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:22 crc kubenswrapper[4813]: E1206 15:48:22.632546 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:23.132531728 +0000 UTC m=+143.023411304 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:22 crc kubenswrapper[4813]: I1206 15:48:22.632706 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:22 crc kubenswrapper[4813]: E1206 15:48:22.633032 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:23.133011011 +0000 UTC m=+143.023890577 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:22 crc kubenswrapper[4813]: I1206 15:48:22.709070 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-2twln"] Dec 06 15:48:22 crc kubenswrapper[4813]: I1206 15:48:22.733370 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:22 crc kubenswrapper[4813]: E1206 15:48:22.733611 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:23.233597384 +0000 UTC m=+143.124476960 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:22 crc kubenswrapper[4813]: I1206 15:48:22.824210 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fq7x6"] Dec 06 15:48:22 crc kubenswrapper[4813]: I1206 15:48:22.850689 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:22 crc kubenswrapper[4813]: E1206 15:48:22.851488 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:23.351472904 +0000 UTC m=+143.242352490 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:22 crc kubenswrapper[4813]: I1206 15:48:22.952929 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:22 crc kubenswrapper[4813]: E1206 15:48:22.953540 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:23.453525135 +0000 UTC m=+143.344404711 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.054002 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:23 crc kubenswrapper[4813]: E1206 15:48:23.054290 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:23.554265211 +0000 UTC m=+143.445144787 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.162712 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:23 crc kubenswrapper[4813]: E1206 15:48:23.163020 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:23.66300666 +0000 UTC m=+143.553886236 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.251440 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mllt2"] Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.275160 4813 generic.go:334] "Generic (PLEG): container finished" podID="a1353568-e317-4a7d-8765-675821dac72b" containerID="fe3922708f1a2cd22f029fc52f23cd01a2d323d7d6963f54a6f98668b7e0aa7f" exitCode=0 Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.275213 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" event={"ID":"a1353568-e317-4a7d-8765-675821dac72b","Type":"ContainerDied","Data":"fe3922708f1a2cd22f029fc52f23cd01a2d323d7d6963f54a6f98668b7e0aa7f"} Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.277051 4813 generic.go:334] "Generic (PLEG): container finished" podID="4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a" containerID="940d679860f73ada396292f3b7f8b00a91972c234677ca4cd59fc30c27af7c57" exitCode=0 Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.277084 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" event={"ID":"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a","Type":"ContainerDied","Data":"940d679860f73ada396292f3b7f8b00a91972c234677ca4cd59fc30c27af7c57"} Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.278569 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktnmq" event={"ID":"8f6312e4-347f-4e82-81b9-4cb1aa530777","Type":"ContainerStarted","Data":"147e4ec5f47b036e0cc4cd4b540033b5a4f0af25e29a01a89e3854c54d570cb2"} Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.279199 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jqrv2" event={"ID":"c5ca581c-83f3-472a-962e-2edaf6e3eba9","Type":"ContainerStarted","Data":"ec278adfaec0dcdff38465de5702bb41144c0fb34e23b63d29ab46cb75c7018a"} Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.279852 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nzc7" event={"ID":"8a7d392b-a4af-4e95-b50e-3e45d8c01037","Type":"ContainerStarted","Data":"7416a8c2ec5484ff64fef3c5a4eebb1b40da51846342a86e4541dc05fe958fcc"} Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.280740 4813 generic.go:334] "Generic (PLEG): container finished" podID="8c420468-f370-44cb-9e39-7ae8f12b461e" containerID="528ccf2fab1e938700253d75c734076782f20773fe314316b056bdc241e79687" exitCode=0 Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.280771 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rtsv4" event={"ID":"8c420468-f370-44cb-9e39-7ae8f12b461e","Type":"ContainerDied","Data":"528ccf2fab1e938700253d75c734076782f20773fe314316b056bdc241e79687"} Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.282130 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pmr8t" event={"ID":"30e8f6c3-6c7e-45dc-9a6f-e2de13d1322c","Type":"ContainerStarted","Data":"b187299b387fa56a7d7e666f5b238cfbfd26880e337f738f35dfd81561c276be"} Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.282969 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxx9w" event={"ID":"64703472-c8cf-4b3d-b464-a2ef7eb89c64","Type":"ContainerStarted","Data":"cca8601cf02e8db2ded75ed43988d9e12e05eda319d7a86771dc1f877e14778a"} Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.283519 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d4h8p" event={"ID":"9549da48-ad83-4eff-878d-c9df63e66cd7","Type":"ContainerStarted","Data":"08697d63bd1a608e9ea10d7e18152da686533852474c3852008191e62521552f"} Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.284773 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-dq6hp" event={"ID":"0db3a3aa-f872-43c3-ade2-58c55a448fe3","Type":"ContainerStarted","Data":"e02dac468fc05d703e2276967c165382e6bb5185bd52c5e1a4891c82155567c5"} Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.284792 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-6bbs5" Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.309722 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:23 crc kubenswrapper[4813]: E1206 15:48:23.310009 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:23.80999601 +0000 UTC m=+143.700875586 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.353914 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-wfhsc"] Dec 06 15:48:23 crc kubenswrapper[4813]: W1206 15:48:23.372474 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc522b818_1d20_4b66_afd3_098ebec69e86.slice/crio-b1f516cc19cfc8dd5209bfb40cb3c7f96fcd62df39149269866038c9fdd8b51f WatchSource:0}: Error finding container b1f516cc19cfc8dd5209bfb40cb3c7f96fcd62df39149269866038c9fdd8b51f: Status 404 returned error can't find the container with id b1f516cc19cfc8dd5209bfb40cb3c7f96fcd62df39149269866038c9fdd8b51f Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.374150 4813 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-6bbs5 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.374188 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-6bbs5" podUID="1d5678dc-7a74-4fff-818c-e34321037686" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.453063 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:23 crc kubenswrapper[4813]: E1206 15:48:23.454398 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:23.95437001 +0000 UTC m=+143.845249586 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.518122 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-85tkp"] Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.555725 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:23 crc kubenswrapper[4813]: E1206 15:48:23.556066 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:24.056054242 +0000 UTC m=+143.946933818 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.657341 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:23 crc kubenswrapper[4813]: E1206 15:48:23.657800 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:24.157785605 +0000 UTC m=+144.048665181 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.732903 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pmr8t" podStartSLOduration=124.732886122 podStartE2EDuration="2m4.732886122s" podCreationTimestamp="2025-12-06 15:46:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:23.731885686 +0000 UTC m=+143.622765262" watchObservedRunningTime="2025-12-06 15:48:23.732886122 +0000 UTC m=+143.623765698" Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.734760 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-6bbs5" podStartSLOduration=123.734753872 podStartE2EDuration="2m3.734753872s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:23.712789561 +0000 UTC m=+143.603669127" watchObservedRunningTime="2025-12-06 15:48:23.734753872 +0000 UTC m=+143.625633448" Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.760107 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:23 crc kubenswrapper[4813]: E1206 15:48:23.760512 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:24.260496413 +0000 UTC m=+144.151375989 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.807299 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-gb6cc" podStartSLOduration=123.807284652 podStartE2EDuration="2m3.807284652s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:23.770129628 +0000 UTC m=+143.661009204" watchObservedRunningTime="2025-12-06 15:48:23.807284652 +0000 UTC m=+143.698164228" Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.862188 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:23 crc kubenswrapper[4813]: E1206 15:48:23.862646 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:24.362628167 +0000 UTC m=+144.253507743 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:23 crc kubenswrapper[4813]: I1206 15:48:23.964006 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:23 crc kubenswrapper[4813]: E1206 15:48:23.964502 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:24.464486403 +0000 UTC m=+144.355365979 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.065512 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:24 crc kubenswrapper[4813]: E1206 15:48:24.065902 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:24.565885677 +0000 UTC m=+144.456765253 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.168379 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:24 crc kubenswrapper[4813]: E1206 15:48:24.168676 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:24.668665047 +0000 UTC m=+144.559544623 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.269329 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:24 crc kubenswrapper[4813]: E1206 15:48:24.269832 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:24.769811484 +0000 UTC m=+144.660691060 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.293509 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2twln" event={"ID":"1883cf65-656c-439d-980d-597c40899cc5","Type":"ContainerStarted","Data":"a97feed0c09ce6da00be69517edce5ab02bb73fda7fcadd390aa2b775a6acf4d"} Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.294286 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fq7x6" event={"ID":"231076de-b56c-41ad-830b-8c6699d224ad","Type":"ContainerStarted","Data":"254d9929e7e64d4577fbbd53f9925c10d8dc20adfe332553cd60dc86afa0c472"} Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.295531 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-wfhsc" event={"ID":"45d5fd66-73f2-4eaa-88ad-4cc6bc597387","Type":"ContainerStarted","Data":"9314a51a3fd29c628a069da27bf54c194ebcce622f41371faae44562b73e52ed"} Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.300984 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-dq6hp" event={"ID":"0db3a3aa-f872-43c3-ade2-58c55a448fe3","Type":"ContainerStarted","Data":"89beadac032cbf89819e413f27cb86ddb19f0fd121f6b8fd47d49460267d381a"} Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.302632 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-kz9wk" event={"ID":"c522b818-1d20-4b66-afd3-098ebec69e86","Type":"ContainerStarted","Data":"b1f516cc19cfc8dd5209bfb40cb3c7f96fcd62df39149269866038c9fdd8b51f"} Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.314607 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzb98" event={"ID":"4674583b-142e-4968-9824-7090cfebca75","Type":"ContainerStarted","Data":"e1a5415baa6c9f344ff21277835dffb68e5f9b2f4d0a61c648d7fea33e92cffa"} Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.318777 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-klqt2" event={"ID":"2f02af08-d7ad-4460-b103-cf44d7598449","Type":"ContainerStarted","Data":"b3b7fc87c8eca4530c45ef9b0cfcf1080d008cfb8f1fdfa7e72847d875155183"} Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.319907 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-klqt2" Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.326964 4813 patch_prober.go:28] interesting pod/console-operator-58897d9998-klqt2 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.327047 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-klqt2" podUID="2f02af08-d7ad-4460-b103-cf44d7598449" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.335830 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-dq6hp" podStartSLOduration=124.335813561 podStartE2EDuration="2m4.335813561s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:24.327892842 +0000 UTC m=+144.218772418" watchObservedRunningTime="2025-12-06 15:48:24.335813561 +0000 UTC m=+144.226693137" Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.340646 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" event={"ID":"994c44ca-f520-4452-b267-961bd400908e","Type":"ContainerStarted","Data":"032b09a0f6c8db36ef202f022f145601133f8c4a49b4551eec4e58d2758c9caa"} Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.341506 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.344553 4813 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-fszk7 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.8:6443/healthz\": dial tcp 10.217.0.8:6443: connect: connection refused" start-of-body= Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.344667 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" podUID="994c44ca-f520-4452-b267-961bd400908e" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.8:6443/healthz\": dial tcp 10.217.0.8:6443: connect: connection refused" Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.350618 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mllt2" event={"ID":"935ac385-32ee-46c5-bffb-69b89b4af6a9","Type":"ContainerStarted","Data":"9163b303049f939809b0c6b5a0b1e5959422b820efb136ec8c1aa828d5cb20b4"} Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.360356 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-rkmnv"] Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.374946 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:24 crc kubenswrapper[4813]: E1206 15:48:24.375348 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:24.875325397 +0000 UTC m=+144.766204973 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.376638 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-c7cx6"] Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.387600 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-klqt2" podStartSLOduration=125.387582372 podStartE2EDuration="2m5.387582372s" podCreationTimestamp="2025-12-06 15:46:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:24.386724509 +0000 UTC m=+144.277604085" watchObservedRunningTime="2025-12-06 15:48:24.387582372 +0000 UTC m=+144.278461948" Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.394664 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-85tkp" event={"ID":"9a238598-92cd-4641-ae82-6e471740ee7f","Type":"ContainerStarted","Data":"0be668ca80357db1b5dc24c385624c1b2704296c70a80a8c92b5835392098408"} Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.409813 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-5t7m6" event={"ID":"d57f80ec-52ac-409d-9dc1-93a89d59f714","Type":"ContainerStarted","Data":"c7ccfbcadbef161c97bd15bf18386d26a4fb9a8255308658ca1d0d6932951d12"} Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.463900 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxx9w" event={"ID":"64703472-c8cf-4b3d-b464-a2ef7eb89c64","Type":"ContainerStarted","Data":"3380d2216366de8c6c5e2748250946a69154a8ddd8e4589b3e9dafb31bd33c19"} Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.479660 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:24 crc kubenswrapper[4813]: E1206 15:48:24.480603 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:24.980579813 +0000 UTC m=+144.871459449 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.485411 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" podStartSLOduration=125.48539156 podStartE2EDuration="2m5.48539156s" podCreationTimestamp="2025-12-06 15:46:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:24.479676039 +0000 UTC m=+144.370555615" watchObservedRunningTime="2025-12-06 15:48:24.48539156 +0000 UTC m=+144.376271136" Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.519568 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q" event={"ID":"4b409627-4716-4bdd-a660-4f88693150a2","Type":"ContainerStarted","Data":"2f140a4a1f10630431e04b37c794615c8c1ed29d7b4d0ce308142240350cea09"} Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.519892 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q" Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.522910 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-ncvs7" event={"ID":"8ddf8368-8232-4211-bfca-642b1acef6a4","Type":"ContainerStarted","Data":"b0ff3d52031e7e95956ebf6623eb945495606b277801a6e2b11da8d6e919da6e"} Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.529229 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxx9w" podStartSLOduration=124.52921409 podStartE2EDuration="2m4.52921409s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:24.524002402 +0000 UTC m=+144.414881978" watchObservedRunningTime="2025-12-06 15:48:24.52921409 +0000 UTC m=+144.420093666" Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.537183 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-6bbs5" Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.541205 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-dld8t"] Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.563311 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-5t7m6" podStartSLOduration=125.563294862 podStartE2EDuration="2m5.563294862s" podCreationTimestamp="2025-12-06 15:46:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:24.556495612 +0000 UTC m=+144.447375188" watchObservedRunningTime="2025-12-06 15:48:24.563294862 +0000 UTC m=+144.454174438" Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.585049 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:24 crc kubenswrapper[4813]: E1206 15:48:24.586288 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:25.08626069 +0000 UTC m=+144.977140266 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.657051 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q" podStartSLOduration=124.657032484 podStartE2EDuration="2m4.657032484s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:24.600523218 +0000 UTC m=+144.491402794" watchObservedRunningTime="2025-12-06 15:48:24.657032484 +0000 UTC m=+144.547912060" Dec 06 15:48:24 crc kubenswrapper[4813]: W1206 15:48:24.687419 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode16c161d_a64d_4bc9_a223_aaf7bfc53083.slice/crio-5715478d2cc1072622767422f582d050c2b77a24d2fea12f1f36c15e4b416a67 WatchSource:0}: Error finding container 5715478d2cc1072622767422f582d050c2b77a24d2fea12f1f36c15e4b416a67: Status 404 returned error can't find the container with id 5715478d2cc1072622767422f582d050c2b77a24d2fea12f1f36c15e4b416a67 Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.688095 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:24 crc kubenswrapper[4813]: E1206 15:48:24.690062 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:25.190040357 +0000 UTC m=+145.080919933 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.791405 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:24 crc kubenswrapper[4813]: E1206 15:48:24.791777 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:25.29176548 +0000 UTC m=+145.182645056 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.892779 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:24 crc kubenswrapper[4813]: E1206 15:48:24.893305 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:25.393264636 +0000 UTC m=+145.284144212 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:24 crc kubenswrapper[4813]: I1206 15:48:24.995093 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:24 crc kubenswrapper[4813]: E1206 15:48:24.995604 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:25.495591415 +0000 UTC m=+145.386470991 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.096843 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:25 crc kubenswrapper[4813]: E1206 15:48:25.097566 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:25.597549194 +0000 UTC m=+145.488428770 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.136379 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-mwqcj"] Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.152412 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-t8zwg"] Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.202345 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:25 crc kubenswrapper[4813]: E1206 15:48:25.202660 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:25.702648075 +0000 UTC m=+145.593527641 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.304057 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:25 crc kubenswrapper[4813]: E1206 15:48:25.304665 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:25.804648665 +0000 UTC m=+145.695528241 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.360710 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cbbdg"] Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.408656 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:25 crc kubenswrapper[4813]: E1206 15:48:25.409116 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:25.90909971 +0000 UTC m=+145.799979286 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.412625 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-kkbqh"] Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.438808 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q" Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.466519 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417265-hcswl"] Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.498810 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x54c"] Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.533545 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:25 crc kubenswrapper[4813]: E1206 15:48:25.534328 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:26.034312234 +0000 UTC m=+145.925191810 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.630562 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jqrv2" event={"ID":"c5ca581c-83f3-472a-962e-2edaf6e3eba9","Type":"ContainerStarted","Data":"e4cba8c1c53ad7f725558c5731f0e3ed7576aa114959d6e188d3fdbbb3f4ca31"} Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.631889 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jqrv2" Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.634998 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:25 crc kubenswrapper[4813]: E1206 15:48:25.635321 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:26.135309257 +0000 UTC m=+146.026188833 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.644711 4813 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-jqrv2 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.22:5443/healthz\": dial tcp 10.217.0.22:5443: connect: connection refused" start-of-body= Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.644756 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jqrv2" podUID="c5ca581c-83f3-472a-962e-2edaf6e3eba9" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.22:5443/healthz\": dial tcp 10.217.0.22:5443: connect: connection refused" Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.677176 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-85tkp" event={"ID":"9a238598-92cd-4641-ae82-6e471740ee7f","Type":"ContainerStarted","Data":"12dac658b651ecfd58cced7e18dde7146e4158762b5faec71036efb54788f51f"} Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.725594 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-ncvs7" event={"ID":"8ddf8368-8232-4211-bfca-642b1acef6a4","Type":"ContainerStarted","Data":"694f9eb7647647f82d8a67124177fe171b5a1dc4b80160ad0f892ff6274125c7"} Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.736687 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:25 crc kubenswrapper[4813]: E1206 15:48:25.737840 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:26.237822161 +0000 UTC m=+146.128701737 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.769877 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bzqcl"] Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.770068 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c7cx6" event={"ID":"f2dde1d1-6ebd-4592-a0cf-9a61c8b7b4ae","Type":"ContainerStarted","Data":"f5c3cbffe5ba5fcc6fb212538a2f9a4657dfafcd73af866bdce9efd7374825e8"} Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.802025 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jqrv2" podStartSLOduration=125.80201131 podStartE2EDuration="2m5.80201131s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:25.801286061 +0000 UTC m=+145.692165637" watchObservedRunningTime="2025-12-06 15:48:25.80201131 +0000 UTC m=+145.692890886" Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.842731 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:25 crc kubenswrapper[4813]: E1206 15:48:25.843773 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:26.343762565 +0000 UTC m=+146.234642141 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.884993 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fq7x6" event={"ID":"231076de-b56c-41ad-830b-8c6699d224ad","Type":"ContainerStarted","Data":"b9b86a5ded46d79153c14e4c9a60e909a6cd6c82fdc1361dc461d80ef78cf65d"} Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.885042 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fq7x6" event={"ID":"231076de-b56c-41ad-830b-8c6699d224ad","Type":"ContainerStarted","Data":"77c69526c691b88592d8e588cf935a192a2f9c08928d9426d06ff950cbd2a206"} Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.885840 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fq7x6" Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.912006 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-t8zwg" event={"ID":"2e8f4e1f-b111-4e8f-8600-6ac82ded86e9","Type":"ContainerStarted","Data":"85e7b4be28342e194ddcf0475d5cd706ea21cdca055f5e17a7c1f978d549e3c7"} Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.912751 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-85tkp" podStartSLOduration=125.912732821 podStartE2EDuration="2m5.912732821s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:25.91156882 +0000 UTC m=+145.802448386" watchObservedRunningTime="2025-12-06 15:48:25.912732821 +0000 UTC m=+145.803612397" Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.944610 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:25 crc kubenswrapper[4813]: E1206 15:48:25.944971 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:26.444950393 +0000 UTC m=+146.335829969 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:25 crc kubenswrapper[4813]: I1206 15:48:25.967159 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mllt2" event={"ID":"935ac385-32ee-46c5-bffb-69b89b4af6a9","Type":"ContainerStarted","Data":"7b65786840de004f2439754ec9dce90348b376f705f744a07398b4b520149d0f"} Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.006841 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" event={"ID":"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a","Type":"ContainerStarted","Data":"427d467ed74e3b45675edc3cb8210c07b3a8cabeeafadc411a4de3b93b3502f1"} Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.033900 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktnmq" event={"ID":"8f6312e4-347f-4e82-81b9-4cb1aa530777","Type":"ContainerStarted","Data":"3fd9449a5d7761f6a059bd2027cb289915b0719e5b07feb1fe8fb4177a500298"} Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.045595 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:26 crc kubenswrapper[4813]: E1206 15:48:26.046396 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:26.546383048 +0000 UTC m=+146.437262624 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.070907 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d4h8p" event={"ID":"9549da48-ad83-4eff-878d-c9df63e66cd7","Type":"ContainerStarted","Data":"74a8843fb83e5efd97d8dc07e47a8900eddf55b0f9e207f39cf5babd5f3be72c"} Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.098720 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2twln" event={"ID":"1883cf65-656c-439d-980d-597c40899cc5","Type":"ContainerStarted","Data":"0ae95a8a844fb72435369422e19e786851158b8d5a4f2adf5dd4da40ded42ee1"} Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.114612 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tw7m2"] Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.116662 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-ncvs7" podStartSLOduration=126.116631267 podStartE2EDuration="2m6.116631267s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:26.098534088 +0000 UTC m=+145.989413664" watchObservedRunningTime="2025-12-06 15:48:26.116631267 +0000 UTC m=+146.007510843" Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.124827 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-wj67k"] Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.141950 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rtsv4" event={"ID":"8c420468-f370-44cb-9e39-7ae8f12b461e","Type":"ContainerStarted","Data":"c05b2bada8f0954069ca0c026f9a289c651860a3c02d53acfac53b4dc62467a3"} Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.141982 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rtsv4" Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.157408 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:26 crc kubenswrapper[4813]: E1206 15:48:26.158202 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:26.658168417 +0000 UTC m=+146.549047993 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.160833 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-dld8t" event={"ID":"279e4882-ef42-445a-891f-65dff6bdd8dc","Type":"ContainerStarted","Data":"b2e8a64c2fc75f64abbfd1575ccf2e3aeda766e58e0ee2841cdecf3738270823"} Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.163438 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-tm2zm"] Dec 06 15:48:26 crc kubenswrapper[4813]: W1206 15:48:26.172322 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0904de59_781e_4c00_86bd_ccebf7a06587.slice/crio-2a05e1c4de263d861117980a637ba24ef210a3cd7bfe344b737c73fd00d1b65c WatchSource:0}: Error finding container 2a05e1c4de263d861117980a637ba24ef210a3cd7bfe344b737c73fd00d1b65c: Status 404 returned error can't find the container with id 2a05e1c4de263d861117980a637ba24ef210a3cd7bfe344b737c73fd00d1b65c Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.174825 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nzc7" event={"ID":"8a7d392b-a4af-4e95-b50e-3e45d8c01037","Type":"ContainerStarted","Data":"3f8d060a373b1d5345f6d0fb92ff222466325f80e51782c515abd145291ae63a"} Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.200966 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-mwqcj" event={"ID":"3b977b1d-5f16-4989-b851-145fe091af2e","Type":"ContainerStarted","Data":"c273ca8f856d5af302f413ae2d868de10c5e98ba46f416cd8b6103d0c863be32"} Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.206081 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mllt2" podStartSLOduration=126.206062875 podStartE2EDuration="2m6.206062875s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:26.201124844 +0000 UTC m=+146.092004420" watchObservedRunningTime="2025-12-06 15:48:26.206062875 +0000 UTC m=+146.096942471" Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.226996 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzb98" event={"ID":"4674583b-142e-4968-9824-7090cfebca75","Type":"ContainerStarted","Data":"ec3f3436e73e50c17e0d46fa2a12b675561f74a5e06689bb8f37fa945072a761"} Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.247116 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" event={"ID":"a1353568-e317-4a7d-8765-675821dac72b","Type":"ContainerStarted","Data":"a9267075176dc10e5c884705f89c2f9ec93cf6d720973dc4a69fe56cf21b2b11"} Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.260251 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d4h8p" podStartSLOduration=127.260230368 podStartE2EDuration="2m7.260230368s" podCreationTimestamp="2025-12-06 15:46:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:26.260185327 +0000 UTC m=+146.151064903" watchObservedRunningTime="2025-12-06 15:48:26.260230368 +0000 UTC m=+146.151109954" Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.261022 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:26 crc kubenswrapper[4813]: E1206 15:48:26.272122 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:26.772102183 +0000 UTC m=+146.662981759 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.288292 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rkmnv" event={"ID":"e16c161d-a64d-4bc9-a223-aaf7bfc53083","Type":"ContainerStarted","Data":"e67c2e9e36f43cc5a72048ea6ab68b9fa87527c383d345085d54ec7094443463"} Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.288341 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rkmnv" event={"ID":"e16c161d-a64d-4bc9-a223-aaf7bfc53083","Type":"ContainerStarted","Data":"5715478d2cc1072622767422f582d050c2b77a24d2fea12f1f36c15e4b416a67"} Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.310982 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-78nhb"] Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.349429 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-ncvs7" Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.352552 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-kz9wk" event={"ID":"c522b818-1d20-4b66-afd3-098ebec69e86","Type":"ContainerStarted","Data":"9341fa3a9a21a64300a80255d3068dcf7eded3760847edf3b20d70a616be6b1c"} Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.369603 4813 patch_prober.go:28] interesting pod/router-default-5444994796-ncvs7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 15:48:26 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 06 15:48:26 crc kubenswrapper[4813]: [+]process-running ok Dec 06 15:48:26 crc kubenswrapper[4813]: healthz check failed Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.369640 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncvs7" podUID="8ddf8368-8232-4211-bfca-642b1acef6a4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.369704 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.370417 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fq7x6" podStartSLOduration=126.370408355 podStartE2EDuration="2m6.370408355s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:26.369965053 +0000 UTC m=+146.260844629" watchObservedRunningTime="2025-12-06 15:48:26.370408355 +0000 UTC m=+146.261287931" Dec 06 15:48:26 crc kubenswrapper[4813]: E1206 15:48:26.370744 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:26.870729063 +0000 UTC m=+146.761608639 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.389182 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-wfhsc" event={"ID":"45d5fd66-73f2-4eaa-88ad-4cc6bc597387","Type":"ContainerStarted","Data":"ab1966065faca53e12937b499e880b6aaf157371d885d4aef9bc62c5a189ce4e"} Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.474013 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:26 crc kubenswrapper[4813]: E1206 15:48:26.475135 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:26.975122846 +0000 UTC m=+146.866002422 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.480594 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-289lx"] Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.534602 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-klqt2" Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.560338 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ktnmq" podStartSLOduration=127.560322652 podStartE2EDuration="2m7.560322652s" podCreationTimestamp="2025-12-06 15:46:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:26.483440266 +0000 UTC m=+146.374319842" watchObservedRunningTime="2025-12-06 15:48:26.560322652 +0000 UTC m=+146.451202228" Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.575934 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:26 crc kubenswrapper[4813]: E1206 15:48:26.577098 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:27.077078945 +0000 UTC m=+146.967958521 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.578694 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:26 crc kubenswrapper[4813]: E1206 15:48:26.580356 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:27.080344231 +0000 UTC m=+146.971223807 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.637241 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rtsv4" podStartSLOduration=127.637222527 podStartE2EDuration="2m7.637222527s" podCreationTimestamp="2025-12-06 15:46:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:26.561673047 +0000 UTC m=+146.452552623" watchObservedRunningTime="2025-12-06 15:48:26.637222527 +0000 UTC m=+146.528102103" Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.685815 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:26 crc kubenswrapper[4813]: E1206 15:48:26.686183 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:27.186164462 +0000 UTC m=+147.077044038 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.730215 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-wfhsc" podStartSLOduration=126.730199298 podStartE2EDuration="2m6.730199298s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:26.638975873 +0000 UTC m=+146.529855449" watchObservedRunningTime="2025-12-06 15:48:26.730199298 +0000 UTC m=+146.621078874" Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.787552 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:26 crc kubenswrapper[4813]: E1206 15:48:26.788035 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:27.288023208 +0000 UTC m=+147.178902784 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:26 crc kubenswrapper[4813]: W1206 15:48:26.817475 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60b3916a_79fc_42b6_86c6_4cae051457ca.slice/crio-0437e6a102086e57333e6f9f9d3885373f64e275fc8466ca2c7ad01c94bd06f1 WatchSource:0}: Error finding container 0437e6a102086e57333e6f9f9d3885373f64e275fc8466ca2c7ad01c94bd06f1: Status 404 returned error can't find the container with id 0437e6a102086e57333e6f9f9d3885373f64e275fc8466ca2c7ad01c94bd06f1 Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.821870 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2twln" podStartSLOduration=126.821850884 podStartE2EDuration="2m6.821850884s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:26.772770485 +0000 UTC m=+146.663650061" watchObservedRunningTime="2025-12-06 15:48:26.821850884 +0000 UTC m=+146.712730460" Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.824043 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-5cbx6"] Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.880880 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nzc7" podStartSLOduration=126.880866156 podStartE2EDuration="2m6.880866156s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:26.876370667 +0000 UTC m=+146.767250243" watchObservedRunningTime="2025-12-06 15:48:26.880866156 +0000 UTC m=+146.771745732" Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.885088 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-nvs5q"] Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.888486 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:26 crc kubenswrapper[4813]: E1206 15:48:26.888774 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:27.388760065 +0000 UTC m=+147.279639641 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.950697 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-kz9wk" podStartSLOduration=8.950679234 podStartE2EDuration="8.950679234s" podCreationTimestamp="2025-12-06 15:48:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:26.94071237 +0000 UTC m=+146.831591956" watchObservedRunningTime="2025-12-06 15:48:26.950679234 +0000 UTC m=+146.841558810" Dec 06 15:48:26 crc kubenswrapper[4813]: I1206 15:48:26.989990 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:26 crc kubenswrapper[4813]: E1206 15:48:26.990498 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:27.490484837 +0000 UTC m=+147.381364413 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:27 crc kubenswrapper[4813]: W1206 15:48:27.048060 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b0e59fb_3d78_4720_bcc2_2c2eb985b7bf.slice/crio-d9b57b1828f02cd93e6183a43efbdeb9b887cc26e1aa712047e445b629119351 WatchSource:0}: Error finding container d9b57b1828f02cd93e6183a43efbdeb9b887cc26e1aa712047e445b629119351: Status 404 returned error can't find the container with id d9b57b1828f02cd93e6183a43efbdeb9b887cc26e1aa712047e445b629119351 Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.058247 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" podStartSLOduration=127.058227369 podStartE2EDuration="2m7.058227369s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:27.048336488 +0000 UTC m=+146.939216064" watchObservedRunningTime="2025-12-06 15:48:27.058227369 +0000 UTC m=+146.949106945" Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.092353 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:27 crc kubenswrapper[4813]: E1206 15:48:27.093074 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:27.593057451 +0000 UTC m=+147.483937027 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.130650 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-dld8t" podStartSLOduration=9.130632116 podStartE2EDuration="9.130632116s" podCreationTimestamp="2025-12-06 15:48:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:27.129778803 +0000 UTC m=+147.020658389" watchObservedRunningTime="2025-12-06 15:48:27.130632116 +0000 UTC m=+147.021511692" Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.202646 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:27 crc kubenswrapper[4813]: E1206 15:48:27.202964 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:27.70295326 +0000 UTC m=+147.593832836 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.287725 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rkmnv" podStartSLOduration=127.287709313 podStartE2EDuration="2m7.287709313s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:27.203180506 +0000 UTC m=+147.094060082" watchObservedRunningTime="2025-12-06 15:48:27.287709313 +0000 UTC m=+147.178588889" Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.304136 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:27 crc kubenswrapper[4813]: E1206 15:48:27.304509 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:27.804488958 +0000 UTC m=+147.695368534 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.362453 4813 patch_prober.go:28] interesting pod/router-default-5444994796-ncvs7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 15:48:27 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 06 15:48:27 crc kubenswrapper[4813]: [+]process-running ok Dec 06 15:48:27 crc kubenswrapper[4813]: healthz check failed Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.362512 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncvs7" podUID="8ddf8368-8232-4211-bfca-642b1acef6a4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.386310 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lzb98" podStartSLOduration=127.386292163 podStartE2EDuration="2m7.386292163s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:27.288773902 +0000 UTC m=+147.179653468" watchObservedRunningTime="2025-12-06 15:48:27.386292163 +0000 UTC m=+147.277171739" Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.391589 4813 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-fszk7 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.8:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.391637 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" podUID="994c44ca-f520-4452-b267-961bd400908e" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.8:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.405917 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.405971 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.405989 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.406015 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:48:27 crc kubenswrapper[4813]: E1206 15:48:27.406253 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:27.906238411 +0000 UTC m=+147.797117987 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.412402 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.448462 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-78nhb" event={"ID":"bb3d374a-97d3-4cf8-972a-8e8b795eb9b1","Type":"ContainerStarted","Data":"0119389b6a3ea66b0721061ff4d52932d6683fc5c9908db188b5c05a60bd14f2"} Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.468685 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.479055 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.483870 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rkmnv" event={"ID":"e16c161d-a64d-4bc9-a223-aaf7bfc53083","Type":"ContainerStarted","Data":"6e7007965079518b0961011633fbc64d34ada777ae20594c8c702d75f2f04c4e"} Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.506669 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:27 crc kubenswrapper[4813]: E1206 15:48:27.507793 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:28.007776418 +0000 UTC m=+147.898655994 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.566324 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c7cx6" event={"ID":"f2dde1d1-6ebd-4592-a0cf-9a61c8b7b4ae","Type":"ContainerStarted","Data":"21ef83c2070495bd5c3cddeaaa499e750fd9107793fdf3e31c334820e513660b"} Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.566457 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c7cx6" event={"ID":"f2dde1d1-6ebd-4592-a0cf-9a61c8b7b4ae","Type":"ContainerStarted","Data":"7d774109e4452f10d9df842703c334ad5129abfcf783e8a6ca724c443002262b"} Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.597977 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-289lx" event={"ID":"60b3916a-79fc-42b6-86c6-4cae051457ca","Type":"ContainerStarted","Data":"0437e6a102086e57333e6f9f9d3885373f64e275fc8466ca2c7ad01c94bd06f1"} Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.609061 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:27 crc kubenswrapper[4813]: E1206 15:48:27.609857 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:28.10984012 +0000 UTC m=+148.000719696 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.632519 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417265-hcswl" event={"ID":"8810f12c-6dbc-4bf9-b27e-29ebc5986955","Type":"ContainerStarted","Data":"2fba0692193bb5fa545d890a31ff21e880014c4c26c01164df7322b66f8ec130"} Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.632569 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417265-hcswl" event={"ID":"8810f12c-6dbc-4bf9-b27e-29ebc5986955","Type":"ContainerStarted","Data":"7c38aa4dffc07b1b3e5abbe869db360dcb5fa4bace0370a0c752b04a317dce13"} Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.662048 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" event={"ID":"4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a","Type":"ContainerStarted","Data":"fad51ba9e69542c66f967206560b82d092b5ba686f3421f1d6dc9ae0a1f5f6b7"} Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.671467 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c7cx6" podStartSLOduration=127.671454631 podStartE2EDuration="2m7.671454631s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:27.606189913 +0000 UTC m=+147.497069489" watchObservedRunningTime="2025-12-06 15:48:27.671454631 +0000 UTC m=+147.562334207" Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.672791 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29417265-hcswl" podStartSLOduration=127.672785366 podStartE2EDuration="2m7.672785366s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:27.670634219 +0000 UTC m=+147.561513795" watchObservedRunningTime="2025-12-06 15:48:27.672785366 +0000 UTC m=+147.563664942" Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.681700 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bzqcl" event={"ID":"e8aef227-c82b-4e7d-bf13-d99c935ad12b","Type":"ContainerStarted","Data":"10d2dc3299563eb2a17341a9e84a8ccb507a3a930492df361ddf7b1ef957e6d6"} Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.681747 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bzqcl" event={"ID":"e8aef227-c82b-4e7d-bf13-d99c935ad12b","Type":"ContainerStarted","Data":"99349d382518cd9b851f551bfe0af099371bf6536c8419019e25a2696569808e"} Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.682463 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bzqcl" Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.688247 4813 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-bzqcl container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.688299 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bzqcl" podUID="e8aef227-c82b-4e7d-bf13-d99c935ad12b" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.709683 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.709881 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:48:27 crc kubenswrapper[4813]: E1206 15:48:27.710946 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:28.210927045 +0000 UTC m=+148.101806621 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.711079 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2twln" event={"ID":"1883cf65-656c-439d-980d-597c40899cc5","Type":"ContainerStarted","Data":"2a92441dc17047a55fed607639c3b634000905ec159b9cb7278f457c10fa82a3"} Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.711827 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.726690 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.729910 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.730332 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-t8zwg" event={"ID":"2e8f4e1f-b111-4e8f-8600-6ac82ded86e9","Type":"ContainerStarted","Data":"70ab66c55831f51e752d9d2599d0c272ebac0d0523bd2abb03d11bd2fb0c2a94"} Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.744500 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.747660 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5cbx6" event={"ID":"0664e3bf-bbbe-4547-9697-785efea4dd99","Type":"ContainerStarted","Data":"748e4053f59e6dfdd2ce3d1f282b444f5c79257895869a1ce398dc5a05f8c9ef"} Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.796448 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x54c" event={"ID":"bb28f2df-5922-47fc-aaf0-5887ca9eb03c","Type":"ContainerStarted","Data":"5fbb56d5d2b0fb47cf40f51334646a55fcd4392332171d52f424c0e5248b831c"} Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.796823 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x54c" event={"ID":"bb28f2df-5922-47fc-aaf0-5887ca9eb03c","Type":"ContainerStarted","Data":"d6e5836b6870023bc85f4a0634cf145d8d090837b6cb5f760347d080a8d9f9a8"} Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.811007 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:27 crc kubenswrapper[4813]: E1206 15:48:27.812916 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:28.312900745 +0000 UTC m=+148.203780321 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.820143 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cbbdg" event={"ID":"36af9565-0192-4fee-9729-2bb06631e150","Type":"ContainerStarted","Data":"5d51dcf8c10b777b105161e01a48a06f8a159dc9726fc76027fc8781b5cb86a0"} Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.820185 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cbbdg" event={"ID":"36af9565-0192-4fee-9729-2bb06631e150","Type":"ContainerStarted","Data":"e231cbdb3f415191f43fa5b1a81498e7845b014b534ba840de8d98eba8a20dab"} Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.829311 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nvs5q" event={"ID":"8b0e59fb-3d78-4720-bcc2-2c2eb985b7bf","Type":"ContainerStarted","Data":"d9b57b1828f02cd93e6183a43efbdeb9b887cc26e1aa712047e445b629119351"} Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.864801 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tw7m2" event={"ID":"56d1f612-b545-4fe1-97bd-727b3d8bf062","Type":"ContainerStarted","Data":"c030db14823fa7f189f43d40bc604d3c7d3fc3622b81f6267b8d94e768eace0f"} Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.864847 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tw7m2" event={"ID":"56d1f612-b545-4fe1-97bd-727b3d8bf062","Type":"ContainerStarted","Data":"88150391b62ac07eff2a4a528b48dad662c5a0879bddd5dc9cbd617b1911ea30"} Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.865622 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tw7m2" Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.867035 4813 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-tw7m2 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.867088 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tw7m2" podUID="56d1f612-b545-4fe1-97bd-727b3d8bf062" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.870422 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" podStartSLOduration=128.870411417 podStartE2EDuration="2m8.870411417s" podCreationTimestamp="2025-12-06 15:46:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:27.771868999 +0000 UTC m=+147.662748575" watchObservedRunningTime="2025-12-06 15:48:27.870411417 +0000 UTC m=+147.761290983" Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.870705 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-tm2zm" event={"ID":"1e457a6a-af4c-4d2b-a2af-c758de365a40","Type":"ContainerStarted","Data":"69639d4ec9ce1d33dfe8f635d4bf946c1696a11a571df5397f2cba40ea1bfc28"} Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.870803 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-tm2zm" event={"ID":"1e457a6a-af4c-4d2b-a2af-c758de365a40","Type":"ContainerStarted","Data":"0e7fda3702fc481d10163b34a4e607c847393ee7cbafc7703c1f5e96e7d892be"} Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.876423 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wj67k" event={"ID":"0904de59-781e-4c00-86bd-ccebf7a06587","Type":"ContainerStarted","Data":"2a05e1c4de263d861117980a637ba24ef210a3cd7bfe344b737c73fd00d1b65c"} Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.882101 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-dld8t" event={"ID":"279e4882-ef42-445a-891f-65dff6bdd8dc","Type":"ContainerStarted","Data":"9ff1bfde26dc3e9c6df1796e8c600434470687d14894b79af9a184ea13ceaa49"} Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.912005 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:27 crc kubenswrapper[4813]: E1206 15:48:27.913551 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:28.413524998 +0000 UTC m=+148.304404574 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.922710 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-mwqcj" event={"ID":"3b977b1d-5f16-4989-b851-145fe091af2e","Type":"ContainerStarted","Data":"ebfcaf14c37884ad8095d499eddc802a3296a458680cb72356dc8d6d41d0c70d"} Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.951919 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bzqcl" podStartSLOduration=127.951903224 podStartE2EDuration="2m7.951903224s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:27.869746709 +0000 UTC m=+147.760626275" watchObservedRunningTime="2025-12-06 15:48:27.951903224 +0000 UTC m=+147.842782800" Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.952347 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cbbdg" podStartSLOduration=127.952341106 podStartE2EDuration="2m7.952341106s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:27.944706253 +0000 UTC m=+147.835585829" watchObservedRunningTime="2025-12-06 15:48:27.952341106 +0000 UTC m=+147.843220672" Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.962167 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-kkbqh" event={"ID":"26bc2fa2-9eee-4f7b-b31b-86f364df2b06","Type":"ContainerStarted","Data":"f89c7bb5a3731d5694812efe5ca0028f56d9afab1eb2aef5e5d8762e852f402b"} Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.962214 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-kkbqh" event={"ID":"26bc2fa2-9eee-4f7b-b31b-86f364df2b06","Type":"ContainerStarted","Data":"63c0def225296b2bd9332616046f9767b8769651eec6f7484fafeb5af0e6c712"} Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.963098 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-kkbqh" Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.969166 4813 patch_prober.go:28] interesting pod/downloads-7954f5f757-kkbqh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Dec 06 15:48:27 crc kubenswrapper[4813]: I1206 15:48:27.969203 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-kkbqh" podUID="26bc2fa2-9eee-4f7b-b31b-86f364df2b06" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Dec 06 15:48:28 crc kubenswrapper[4813]: I1206 15:48:28.013345 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:28 crc kubenswrapper[4813]: E1206 15:48:28.014837 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:28.514826479 +0000 UTC m=+148.405706055 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:28 crc kubenswrapper[4813]: I1206 15:48:28.017110 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tw7m2" podStartSLOduration=128.017090069 podStartE2EDuration="2m8.017090069s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:28.016638197 +0000 UTC m=+147.907517773" watchObservedRunningTime="2025-12-06 15:48:28.017090069 +0000 UTC m=+147.907969645" Dec 06 15:48:28 crc kubenswrapper[4813]: I1206 15:48:28.037706 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d4h8p" event={"ID":"9549da48-ad83-4eff-878d-c9df63e66cd7","Type":"ContainerStarted","Data":"b13065f01d150ba1b4b8bda7b42338f3507459fef2477d020e93ca63ba0049a0"} Dec 06 15:48:28 crc kubenswrapper[4813]: I1206 15:48:28.042360 4813 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rtsv4 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Dec 06 15:48:28 crc kubenswrapper[4813]: I1206 15:48:28.042404 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rtsv4" podUID="8c420468-f370-44cb-9e39-7ae8f12b461e" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Dec 06 15:48:28 crc kubenswrapper[4813]: I1206 15:48:28.048981 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:48:28 crc kubenswrapper[4813]: I1206 15:48:28.052839 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x54c" podStartSLOduration=128.052820475 podStartE2EDuration="2m8.052820475s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:28.051453499 +0000 UTC m=+147.942333075" watchObservedRunningTime="2025-12-06 15:48:28.052820475 +0000 UTC m=+147.943700051" Dec 06 15:48:28 crc kubenswrapper[4813]: I1206 15:48:28.115029 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:28 crc kubenswrapper[4813]: E1206 15:48:28.116627 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:28.616612704 +0000 UTC m=+148.507492280 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:28 crc kubenswrapper[4813]: I1206 15:48:28.217631 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:28 crc kubenswrapper[4813]: E1206 15:48:28.217993 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:28.717977727 +0000 UTC m=+148.608857303 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:28 crc kubenswrapper[4813]: I1206 15:48:28.318769 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:28 crc kubenswrapper[4813]: E1206 15:48:28.318902 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:28.818878337 +0000 UTC m=+148.709757913 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:28 crc kubenswrapper[4813]: I1206 15:48:28.319092 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:28 crc kubenswrapper[4813]: E1206 15:48:28.319379 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:28.81936756 +0000 UTC m=+148.710247136 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:28 crc kubenswrapper[4813]: I1206 15:48:28.370867 4813 patch_prober.go:28] interesting pod/router-default-5444994796-ncvs7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 15:48:28 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 06 15:48:28 crc kubenswrapper[4813]: [+]process-running ok Dec 06 15:48:28 crc kubenswrapper[4813]: healthz check failed Dec 06 15:48:28 crc kubenswrapper[4813]: I1206 15:48:28.370903 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncvs7" podUID="8ddf8368-8232-4211-bfca-642b1acef6a4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 15:48:28 crc kubenswrapper[4813]: I1206 15:48:28.420984 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:28 crc kubenswrapper[4813]: E1206 15:48:28.421532 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:28.921515844 +0000 UTC m=+148.812395420 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:28 crc kubenswrapper[4813]: I1206 15:48:28.527375 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:28 crc kubenswrapper[4813]: E1206 15:48:28.528158 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:29.028144056 +0000 UTC m=+148.919023632 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:28 crc kubenswrapper[4813]: I1206 15:48:28.632029 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:28 crc kubenswrapper[4813]: E1206 15:48:28.632634 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:29.132618522 +0000 UTC m=+149.023498098 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:28 crc kubenswrapper[4813]: I1206 15:48:28.733474 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:28 crc kubenswrapper[4813]: E1206 15:48:28.733783 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:29.233772199 +0000 UTC m=+149.124651775 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:28 crc kubenswrapper[4813]: I1206 15:48:28.842893 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:28 crc kubenswrapper[4813]: E1206 15:48:28.843067 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:29.343039831 +0000 UTC m=+149.233919407 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:28 crc kubenswrapper[4813]: I1206 15:48:28.843435 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:28 crc kubenswrapper[4813]: E1206 15:48:28.843807 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:29.343792161 +0000 UTC m=+149.234671737 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:28 crc kubenswrapper[4813]: I1206 15:48:28.944723 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:28 crc kubenswrapper[4813]: E1206 15:48:28.944908 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:29.444893257 +0000 UTC m=+149.335772833 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:28 crc kubenswrapper[4813]: I1206 15:48:28.945131 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:28 crc kubenswrapper[4813]: E1206 15:48:28.945404 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:29.44539595 +0000 UTC m=+149.336275526 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.038637 4813 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-jqrv2 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.22:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.038699 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jqrv2" podUID="c5ca581c-83f3-472a-962e-2edaf6e3eba9" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.22:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.042026 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nvs5q" event={"ID":"8b0e59fb-3d78-4720-bcc2-2c2eb985b7bf","Type":"ContainerStarted","Data":"659dff3ba43163b3babbb394416219e65f0af95f26308be14ea5b5be7a563cb7"} Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.043834 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-tm2zm" event={"ID":"1e457a6a-af4c-4d2b-a2af-c758de365a40","Type":"ContainerStarted","Data":"2a8c30470c566803b84a68f2876237d12d648e180f7834280808483723b25bd2"} Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.044164 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-tm2zm" Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.045291 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-t8zwg" event={"ID":"2e8f4e1f-b111-4e8f-8600-6ac82ded86e9","Type":"ContainerStarted","Data":"65073417940b21b91b07bd611cf32c0ec9f5e800ffe180edf45253b277c13950"} Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.045771 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:29 crc kubenswrapper[4813]: E1206 15:48:29.046114 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:29.546098566 +0000 UTC m=+149.436978142 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.047207 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5cbx6" event={"ID":"0664e3bf-bbbe-4547-9697-785efea4dd99","Type":"ContainerStarted","Data":"72bbee160c63b7e916a2111c81b2c9212e103437f26354fff9d2380e5d264d61"} Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.047229 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5cbx6" event={"ID":"0664e3bf-bbbe-4547-9697-785efea4dd99","Type":"ContainerStarted","Data":"fa7b8fdc31be3e7612c00737e9b8aa4910b6acebf5838d87ded302287f05b38b"} Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.048378 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-289lx" event={"ID":"60b3916a-79fc-42b6-86c6-4cae051457ca","Type":"ContainerStarted","Data":"441756359cdf72b3e97cab1b8ce564665e57f14ec2279db9364d1706faa70be5"} Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.048962 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-289lx" Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.049783 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wj67k" event={"ID":"0904de59-781e-4c00-86bd-ccebf7a06587","Type":"ContainerStarted","Data":"d64daabac947a3426c8f4fbc0b1e0f9817d925317eb363e1a62c47ec162b7cb7"} Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.057057 4813 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-289lx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" start-of-body= Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.057109 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-289lx" podUID="60b3916a-79fc-42b6-86c6-4cae051457ca" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.062618 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-78nhb" event={"ID":"bb3d374a-97d3-4cf8-972a-8e8b795eb9b1","Type":"ContainerStarted","Data":"c78c0565c3200480dac8e1459c4a4651aff96c6da0428c5e022f94ed86556071"} Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.076381 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-mwqcj" event={"ID":"3b977b1d-5f16-4989-b851-145fe091af2e","Type":"ContainerStarted","Data":"542d0d519838652e553f06855dd5459ccac4338e7ed4ebb7081663d18a8cc9e4"} Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.079675 4813 patch_prober.go:28] interesting pod/downloads-7954f5f757-kkbqh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.079727 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-kkbqh" podUID="26bc2fa2-9eee-4f7b-b31b-86f364df2b06" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.092463 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tw7m2" Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.117145 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nvs5q" podStartSLOduration=129.117128366 podStartE2EDuration="2m9.117128366s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:29.112253847 +0000 UTC m=+149.003133423" watchObservedRunningTime="2025-12-06 15:48:29.117128366 +0000 UTC m=+149.008007942" Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.117438 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-kkbqh" podStartSLOduration=129.117433694 podStartE2EDuration="2m9.117433694s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:28.147552612 +0000 UTC m=+148.038432188" watchObservedRunningTime="2025-12-06 15:48:29.117433694 +0000 UTC m=+149.008313270" Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.148127 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bzqcl" Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.149445 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:29 crc kubenswrapper[4813]: E1206 15:48:29.151815 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:29.651798834 +0000 UTC m=+149.542678410 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.252444 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5cbx6" podStartSLOduration=129.252425667 podStartE2EDuration="2m9.252425667s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:29.203141733 +0000 UTC m=+149.094021309" watchObservedRunningTime="2025-12-06 15:48:29.252425667 +0000 UTC m=+149.143305243" Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.254821 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:29 crc kubenswrapper[4813]: E1206 15:48:29.255155 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:29.755141439 +0000 UTC m=+149.646021015 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.330040 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-t8zwg" podStartSLOduration=129.330024761 podStartE2EDuration="2m9.330024761s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:29.3277252 +0000 UTC m=+149.218604776" watchObservedRunningTime="2025-12-06 15:48:29.330024761 +0000 UTC m=+149.220904337" Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.363453 4813 patch_prober.go:28] interesting pod/router-default-5444994796-ncvs7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 15:48:29 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 06 15:48:29 crc kubenswrapper[4813]: [+]process-running ok Dec 06 15:48:29 crc kubenswrapper[4813]: healthz check failed Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.363509 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncvs7" podUID="8ddf8368-8232-4211-bfca-642b1acef6a4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.365318 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:29 crc kubenswrapper[4813]: E1206 15:48:29.366042 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:29.866029824 +0000 UTC m=+149.756909390 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.382945 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-tm2zm" podStartSLOduration=11.382930192 podStartE2EDuration="11.382930192s" podCreationTimestamp="2025-12-06 15:48:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:29.380841686 +0000 UTC m=+149.271721272" watchObservedRunningTime="2025-12-06 15:48:29.382930192 +0000 UTC m=+149.273809768" Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.428463 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-289lx" podStartSLOduration=129.428443666 podStartE2EDuration="2m9.428443666s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:29.423081604 +0000 UTC m=+149.313961180" watchObservedRunningTime="2025-12-06 15:48:29.428443666 +0000 UTC m=+149.319323242" Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.469731 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:29 crc kubenswrapper[4813]: E1206 15:48:29.470192 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:29.970171951 +0000 UTC m=+149.861051527 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.494611 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-78nhb" podStartSLOduration=129.494596257 podStartE2EDuration="2m9.494596257s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:29.490997492 +0000 UTC m=+149.381877068" watchObservedRunningTime="2025-12-06 15:48:29.494596257 +0000 UTC m=+149.385475833" Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.543104 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-mwqcj" podStartSLOduration=129.543087531 podStartE2EDuration="2m9.543087531s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:29.540651406 +0000 UTC m=+149.431530982" watchObservedRunningTime="2025-12-06 15:48:29.543087531 +0000 UTC m=+149.433967107" Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.575074 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:29 crc kubenswrapper[4813]: E1206 15:48:29.575578 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:30.07556028 +0000 UTC m=+149.966439856 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.675826 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:29 crc kubenswrapper[4813]: E1206 15:48:29.676122 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:30.176106982 +0000 UTC m=+150.066986558 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:29 crc kubenswrapper[4813]: W1206 15:48:29.746651 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-a5f940a688feb81706f9cb8ba0d5c8246a80a5439aa1621df014c5501a587f5c WatchSource:0}: Error finding container a5f940a688feb81706f9cb8ba0d5c8246a80a5439aa1621df014c5501a587f5c: Status 404 returned error can't find the container with id a5f940a688feb81706f9cb8ba0d5c8246a80a5439aa1621df014c5501a587f5c Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.780374 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:29 crc kubenswrapper[4813]: E1206 15:48:29.780996 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:30.280983758 +0000 UTC m=+150.171863334 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.882686 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:29 crc kubenswrapper[4813]: E1206 15:48:29.883299 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:30.383240764 +0000 UTC m=+150.274120340 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.978151 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rtsv4" Dec 06 15:48:29 crc kubenswrapper[4813]: I1206 15:48:29.984034 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:29 crc kubenswrapper[4813]: E1206 15:48:29.984376 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:30.484362961 +0000 UTC m=+150.375242537 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.078985 4813 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-jqrv2 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.22:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.079055 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jqrv2" podUID="c5ca581c-83f3-472a-962e-2edaf6e3eba9" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.22:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.085474 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:30 crc kubenswrapper[4813]: E1206 15:48:30.086540 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:30.586517455 +0000 UTC m=+150.477397031 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.098568 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wj67k" event={"ID":"0904de59-781e-4c00-86bd-ccebf7a06587","Type":"ContainerStarted","Data":"7e792f3ce93a65acf47492ea24289de186af53aca7772270982767ce7a982fcc"} Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.100067 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"20ed41047c1dc4b35544185c3fce5e45c24c47cf1a642e31d8595868536d98ca"} Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.100112 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"0ca4908ba1a01ec52d1b3c617fa43b27317c466961afa0f75146c0905be11d35"} Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.105727 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"a5f940a688feb81706f9cb8ba0d5c8246a80a5439aa1621df014c5501a587f5c"} Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.112867 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"2a45d28cc07885b435861b939f1b9f24314fe73acc65c4605b955e9440521fda"} Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.112904 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"47befeffbbccd234e3264d4efe8c475931618870217cd20f8c49024544c53379"} Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.114173 4813 patch_prober.go:28] interesting pod/downloads-7954f5f757-kkbqh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.114226 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-kkbqh" podUID="26bc2fa2-9eee-4f7b-b31b-86f364df2b06" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.114906 4813 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-289lx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" start-of-body= Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.114924 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-289lx" podUID="60b3916a-79fc-42b6-86c6-4cae051457ca" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.189037 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:30 crc kubenswrapper[4813]: E1206 15:48:30.190839 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:30.690822276 +0000 UTC m=+150.581701852 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.290446 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:30 crc kubenswrapper[4813]: E1206 15:48:30.292834 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:30.792816275 +0000 UTC m=+150.683695851 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.348151 4813 patch_prober.go:28] interesting pod/router-default-5444994796-ncvs7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 15:48:30 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 06 15:48:30 crc kubenswrapper[4813]: [+]process-running ok Dec 06 15:48:30 crc kubenswrapper[4813]: healthz check failed Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.348222 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncvs7" podUID="8ddf8368-8232-4211-bfca-642b1acef6a4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.392396 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:30 crc kubenswrapper[4813]: E1206 15:48:30.392695 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:30.892684189 +0000 UTC m=+150.783563765 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.483167 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.483230 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.492902 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:30 crc kubenswrapper[4813]: E1206 15:48:30.493071 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:30.993036945 +0000 UTC m=+150.883916521 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.493121 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:30 crc kubenswrapper[4813]: E1206 15:48:30.493433 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:30.993421775 +0000 UTC m=+150.884301351 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.510650 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-gb6cc" Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.511284 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-gb6cc" Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.512660 4813 patch_prober.go:28] interesting pod/console-f9d7485db-gb6cc container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.512692 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-gb6cc" podUID="36f25c5a-3e57-4367-b306-db9661e4f7c9" containerName="console" probeResult="failure" output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.585857 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-54xzm"] Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.586815 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-54xzm" Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.595540 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:30 crc kubenswrapper[4813]: E1206 15:48:30.595741 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:31.095710132 +0000 UTC m=+150.986589718 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.596117 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:30 crc kubenswrapper[4813]: E1206 15:48:30.597287 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:31.097250703 +0000 UTC m=+150.988130289 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.615955 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.634206 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.634244 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.639838 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-54xzm"] Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.649964 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.697120 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.697440 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2fac5b3-c60b-4c96-8841-ae53a24aaed1-catalog-content\") pod \"certified-operators-54xzm\" (UID: \"e2fac5b3-c60b-4c96-8841-ae53a24aaed1\") " pod="openshift-marketplace/certified-operators-54xzm" Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.697476 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2fac5b3-c60b-4c96-8841-ae53a24aaed1-utilities\") pod \"certified-operators-54xzm\" (UID: \"e2fac5b3-c60b-4c96-8841-ae53a24aaed1\") " pod="openshift-marketplace/certified-operators-54xzm" Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.697503 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k59xj\" (UniqueName: \"kubernetes.io/projected/e2fac5b3-c60b-4c96-8841-ae53a24aaed1-kube-api-access-k59xj\") pod \"certified-operators-54xzm\" (UID: \"e2fac5b3-c60b-4c96-8841-ae53a24aaed1\") " pod="openshift-marketplace/certified-operators-54xzm" Dec 06 15:48:30 crc kubenswrapper[4813]: E1206 15:48:30.698133 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:31.198116982 +0000 UTC m=+151.088996558 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.793682 4813 patch_prober.go:28] interesting pod/apiserver-76f77b778f-8vzrm container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 06 15:48:30 crc kubenswrapper[4813]: [+]log ok Dec 06 15:48:30 crc kubenswrapper[4813]: [+]etcd ok Dec 06 15:48:30 crc kubenswrapper[4813]: [-]poststarthook/start-apiserver-admission-initializer failed: reason withheld Dec 06 15:48:30 crc kubenswrapper[4813]: [-]poststarthook/generic-apiserver-start-informers failed: reason withheld Dec 06 15:48:30 crc kubenswrapper[4813]: [+]poststarthook/max-in-flight-filter ok Dec 06 15:48:30 crc kubenswrapper[4813]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 06 15:48:30 crc kubenswrapper[4813]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 06 15:48:30 crc kubenswrapper[4813]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 06 15:48:30 crc kubenswrapper[4813]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Dec 06 15:48:30 crc kubenswrapper[4813]: [+]poststarthook/project.openshift.io-projectcache ok Dec 06 15:48:30 crc kubenswrapper[4813]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 06 15:48:30 crc kubenswrapper[4813]: [+]poststarthook/openshift.io-startinformers ok Dec 06 15:48:30 crc kubenswrapper[4813]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 06 15:48:30 crc kubenswrapper[4813]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 06 15:48:30 crc kubenswrapper[4813]: livez check failed Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.793730 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" podUID="4f1eb4c7-a202-428d-88ed-2d0d4f5bc43a" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 15:48:30 crc kubenswrapper[4813]: E1206 15:48:30.799178 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:31.299162967 +0000 UTC m=+151.190042543 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.798852 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.799385 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2fac5b3-c60b-4c96-8841-ae53a24aaed1-catalog-content\") pod \"certified-operators-54xzm\" (UID: \"e2fac5b3-c60b-4c96-8841-ae53a24aaed1\") " pod="openshift-marketplace/certified-operators-54xzm" Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.799454 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2fac5b3-c60b-4c96-8841-ae53a24aaed1-utilities\") pod \"certified-operators-54xzm\" (UID: \"e2fac5b3-c60b-4c96-8841-ae53a24aaed1\") " pod="openshift-marketplace/certified-operators-54xzm" Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.799474 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k59xj\" (UniqueName: \"kubernetes.io/projected/e2fac5b3-c60b-4c96-8841-ae53a24aaed1-kube-api-access-k59xj\") pod \"certified-operators-54xzm\" (UID: \"e2fac5b3-c60b-4c96-8841-ae53a24aaed1\") " pod="openshift-marketplace/certified-operators-54xzm" Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.799949 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2fac5b3-c60b-4c96-8841-ae53a24aaed1-catalog-content\") pod \"certified-operators-54xzm\" (UID: \"e2fac5b3-c60b-4c96-8841-ae53a24aaed1\") " pod="openshift-marketplace/certified-operators-54xzm" Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.800142 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2fac5b3-c60b-4c96-8841-ae53a24aaed1-utilities\") pod \"certified-operators-54xzm\" (UID: \"e2fac5b3-c60b-4c96-8841-ae53a24aaed1\") " pod="openshift-marketplace/certified-operators-54xzm" Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.803141 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.803670 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.819417 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.824775 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.864668 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.886133 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k59xj\" (UniqueName: \"kubernetes.io/projected/e2fac5b3-c60b-4c96-8841-ae53a24aaed1-kube-api-access-k59xj\") pod \"certified-operators-54xzm\" (UID: \"e2fac5b3-c60b-4c96-8841-ae53a24aaed1\") " pod="openshift-marketplace/certified-operators-54xzm" Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.900820 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.901001 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/977c2e53-21f6-4e10-a539-d57bcb28a6bb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"977c2e53-21f6-4e10-a539-d57bcb28a6bb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.901032 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/977c2e53-21f6-4e10-a539-d57bcb28a6bb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"977c2e53-21f6-4e10-a539-d57bcb28a6bb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 15:48:30 crc kubenswrapper[4813]: E1206 15:48:30.901158 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:31.401143696 +0000 UTC m=+151.292023272 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.907599 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-54xzm" Dec 06 15:48:30 crc kubenswrapper[4813]: I1206 15:48:30.916864 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.002760 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/977c2e53-21f6-4e10-a539-d57bcb28a6bb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"977c2e53-21f6-4e10-a539-d57bcb28a6bb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.002802 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/977c2e53-21f6-4e10-a539-d57bcb28a6bb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"977c2e53-21f6-4e10-a539-d57bcb28a6bb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.002833 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.003527 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/977c2e53-21f6-4e10-a539-d57bcb28a6bb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"977c2e53-21f6-4e10-a539-d57bcb28a6bb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 15:48:31 crc kubenswrapper[4813]: E1206 15:48:31.003932 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:31.503922216 +0000 UTC m=+151.394801792 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.074876 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kp6t2"] Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.075740 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kp6t2" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.083889 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/977c2e53-21f6-4e10-a539-d57bcb28a6bb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"977c2e53-21f6-4e10-a539-d57bcb28a6bb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.107970 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:31 crc kubenswrapper[4813]: E1206 15:48:31.108384 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:31.608368871 +0000 UTC m=+151.499248447 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.117085 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.162897 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wj67k" event={"ID":"0904de59-781e-4c00-86bd-ccebf7a06587","Type":"ContainerStarted","Data":"87070685179c86946a07cfd3c724e6fbdf97d5394ce275ad01a08b61d03c751f"} Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.181658 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"6db8bb71bf11b1306ad332d7703c1f62ac03effc7fdd845376db1d9b2b1f6b5b"} Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.182325 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.183097 4813 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-289lx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" start-of-body= Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.183131 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-289lx" podUID="60b3916a-79fc-42b6-86c6-4cae051457ca" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.209842 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e982bb16-5f82-418b-8c03-8c73d22e010b-catalog-content\") pod \"certified-operators-kp6t2\" (UID: \"e982bb16-5f82-418b-8c03-8c73d22e010b\") " pod="openshift-marketplace/certified-operators-kp6t2" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.209900 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.209946 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpc2w\" (UniqueName: \"kubernetes.io/projected/e982bb16-5f82-418b-8c03-8c73d22e010b-kube-api-access-vpc2w\") pod \"certified-operators-kp6t2\" (UID: \"e982bb16-5f82-418b-8c03-8c73d22e010b\") " pod="openshift-marketplace/certified-operators-kp6t2" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.209985 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e982bb16-5f82-418b-8c03-8c73d22e010b-utilities\") pod \"certified-operators-kp6t2\" (UID: \"e982bb16-5f82-418b-8c03-8c73d22e010b\") " pod="openshift-marketplace/certified-operators-kp6t2" Dec 06 15:48:31 crc kubenswrapper[4813]: E1206 15:48:31.210248 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:31.710237267 +0000 UTC m=+151.601116843 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.211014 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b955p" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.229859 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jqrv2" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.284015 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6rwpr"] Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.285221 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6rwpr" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.311339 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.311564 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpc2w\" (UniqueName: \"kubernetes.io/projected/e982bb16-5f82-418b-8c03-8c73d22e010b-kube-api-access-vpc2w\") pod \"certified-operators-kp6t2\" (UID: \"e982bb16-5f82-418b-8c03-8c73d22e010b\") " pod="openshift-marketplace/certified-operators-kp6t2" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.311652 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e982bb16-5f82-418b-8c03-8c73d22e010b-utilities\") pod \"certified-operators-kp6t2\" (UID: \"e982bb16-5f82-418b-8c03-8c73d22e010b\") " pod="openshift-marketplace/certified-operators-kp6t2" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.311715 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e982bb16-5f82-418b-8c03-8c73d22e010b-catalog-content\") pod \"certified-operators-kp6t2\" (UID: \"e982bb16-5f82-418b-8c03-8c73d22e010b\") " pod="openshift-marketplace/certified-operators-kp6t2" Dec 06 15:48:31 crc kubenswrapper[4813]: E1206 15:48:31.313701 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:31.813674245 +0000 UTC m=+151.704553911 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.315804 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e982bb16-5f82-418b-8c03-8c73d22e010b-utilities\") pod \"certified-operators-kp6t2\" (UID: \"e982bb16-5f82-418b-8c03-8c73d22e010b\") " pod="openshift-marketplace/certified-operators-kp6t2" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.316253 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e982bb16-5f82-418b-8c03-8c73d22e010b-catalog-content\") pod \"certified-operators-kp6t2\" (UID: \"e982bb16-5f82-418b-8c03-8c73d22e010b\") " pod="openshift-marketplace/certified-operators-kp6t2" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.342116 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kp6t2"] Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.348388 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-ncvs7" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.395535 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6rwpr"] Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.395686 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.414071 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76073738-387a-4968-a483-8880e866d1e4-catalog-content\") pod \"community-operators-6rwpr\" (UID: \"76073738-387a-4968-a483-8880e866d1e4\") " pod="openshift-marketplace/community-operators-6rwpr" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.414122 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76073738-387a-4968-a483-8880e866d1e4-utilities\") pod \"community-operators-6rwpr\" (UID: \"76073738-387a-4968-a483-8880e866d1e4\") " pod="openshift-marketplace/community-operators-6rwpr" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.414184 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.414221 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqpfc\" (UniqueName: \"kubernetes.io/projected/76073738-387a-4968-a483-8880e866d1e4-kube-api-access-nqpfc\") pod \"community-operators-6rwpr\" (UID: \"76073738-387a-4968-a483-8880e866d1e4\") " pod="openshift-marketplace/community-operators-6rwpr" Dec 06 15:48:31 crc kubenswrapper[4813]: E1206 15:48:31.414509 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:31.914497764 +0000 UTC m=+151.805377340 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.423789 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpc2w\" (UniqueName: \"kubernetes.io/projected/e982bb16-5f82-418b-8c03-8c73d22e010b-kube-api-access-vpc2w\") pod \"certified-operators-kp6t2\" (UID: \"e982bb16-5f82-418b-8c03-8c73d22e010b\") " pod="openshift-marketplace/certified-operators-kp6t2" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.435729 4813 patch_prober.go:28] interesting pod/downloads-7954f5f757-kkbqh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.435776 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-kkbqh" podUID="26bc2fa2-9eee-4f7b-b31b-86f364df2b06" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.435970 4813 patch_prober.go:28] interesting pod/downloads-7954f5f757-kkbqh container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.435996 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-kkbqh" podUID="26bc2fa2-9eee-4f7b-b31b-86f364df2b06" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.447809 4813 patch_prober.go:28] interesting pod/router-default-5444994796-ncvs7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 15:48:31 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 06 15:48:31 crc kubenswrapper[4813]: [+]process-running ok Dec 06 15:48:31 crc kubenswrapper[4813]: healthz check failed Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.448110 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncvs7" podUID="8ddf8368-8232-4211-bfca-642b1acef6a4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.515770 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.515930 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76073738-387a-4968-a483-8880e866d1e4-catalog-content\") pod \"community-operators-6rwpr\" (UID: \"76073738-387a-4968-a483-8880e866d1e4\") " pod="openshift-marketplace/community-operators-6rwpr" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.515953 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76073738-387a-4968-a483-8880e866d1e4-utilities\") pod \"community-operators-6rwpr\" (UID: \"76073738-387a-4968-a483-8880e866d1e4\") " pod="openshift-marketplace/community-operators-6rwpr" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.516069 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqpfc\" (UniqueName: \"kubernetes.io/projected/76073738-387a-4968-a483-8880e866d1e4-kube-api-access-nqpfc\") pod \"community-operators-6rwpr\" (UID: \"76073738-387a-4968-a483-8880e866d1e4\") " pod="openshift-marketplace/community-operators-6rwpr" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.517035 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qnq74"] Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.518043 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qnq74" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.518087 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76073738-387a-4968-a483-8880e866d1e4-catalog-content\") pod \"community-operators-6rwpr\" (UID: \"76073738-387a-4968-a483-8880e866d1e4\") " pod="openshift-marketplace/community-operators-6rwpr" Dec 06 15:48:31 crc kubenswrapper[4813]: E1206 15:48:31.518215 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:32.018200559 +0000 UTC m=+151.909080135 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.528885 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76073738-387a-4968-a483-8880e866d1e4-utilities\") pod \"community-operators-6rwpr\" (UID: \"76073738-387a-4968-a483-8880e866d1e4\") " pod="openshift-marketplace/community-operators-6rwpr" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.578882 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqpfc\" (UniqueName: \"kubernetes.io/projected/76073738-387a-4968-a483-8880e866d1e4-kube-api-access-nqpfc\") pod \"community-operators-6rwpr\" (UID: \"76073738-387a-4968-a483-8880e866d1e4\") " pod="openshift-marketplace/community-operators-6rwpr" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.597578 4813 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-289lx container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" start-of-body= Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.597620 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-289lx" podUID="60b3916a-79fc-42b6-86c6-4cae051457ca" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.597904 4813 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-289lx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" start-of-body= Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.597920 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-289lx" podUID="60b3916a-79fc-42b6-86c6-4cae051457ca" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.617597 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6rwpr" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.621295 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.621350 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f522807f-660d-4791-95f4-f2d9daa58936-utilities\") pod \"community-operators-qnq74\" (UID: \"f522807f-660d-4791-95f4-f2d9daa58936\") " pod="openshift-marketplace/community-operators-qnq74" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.621419 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59zbh\" (UniqueName: \"kubernetes.io/projected/f522807f-660d-4791-95f4-f2d9daa58936-kube-api-access-59zbh\") pod \"community-operators-qnq74\" (UID: \"f522807f-660d-4791-95f4-f2d9daa58936\") " pod="openshift-marketplace/community-operators-qnq74" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.621452 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f522807f-660d-4791-95f4-f2d9daa58936-catalog-content\") pod \"community-operators-qnq74\" (UID: \"f522807f-660d-4791-95f4-f2d9daa58936\") " pod="openshift-marketplace/community-operators-qnq74" Dec 06 15:48:31 crc kubenswrapper[4813]: E1206 15:48:31.621708 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:32.121696758 +0000 UTC m=+152.012576334 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.631446 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qnq74"] Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.722740 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kp6t2" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.723319 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.723454 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f522807f-660d-4791-95f4-f2d9daa58936-catalog-content\") pod \"community-operators-qnq74\" (UID: \"f522807f-660d-4791-95f4-f2d9daa58936\") " pod="openshift-marketplace/community-operators-qnq74" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.723521 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f522807f-660d-4791-95f4-f2d9daa58936-utilities\") pod \"community-operators-qnq74\" (UID: \"f522807f-660d-4791-95f4-f2d9daa58936\") " pod="openshift-marketplace/community-operators-qnq74" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.723570 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59zbh\" (UniqueName: \"kubernetes.io/projected/f522807f-660d-4791-95f4-f2d9daa58936-kube-api-access-59zbh\") pod \"community-operators-qnq74\" (UID: \"f522807f-660d-4791-95f4-f2d9daa58936\") " pod="openshift-marketplace/community-operators-qnq74" Dec 06 15:48:31 crc kubenswrapper[4813]: E1206 15:48:31.723676 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:32.223643057 +0000 UTC m=+152.114522623 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.723949 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f522807f-660d-4791-95f4-f2d9daa58936-catalog-content\") pod \"community-operators-qnq74\" (UID: \"f522807f-660d-4791-95f4-f2d9daa58936\") " pod="openshift-marketplace/community-operators-qnq74" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.724221 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f522807f-660d-4791-95f4-f2d9daa58936-utilities\") pod \"community-operators-qnq74\" (UID: \"f522807f-660d-4791-95f4-f2d9daa58936\") " pod="openshift-marketplace/community-operators-qnq74" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.828618 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:31 crc kubenswrapper[4813]: E1206 15:48:31.829324 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:32.329308463 +0000 UTC m=+152.220188039 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.854310 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59zbh\" (UniqueName: \"kubernetes.io/projected/f522807f-660d-4791-95f4-f2d9daa58936-kube-api-access-59zbh\") pod \"community-operators-qnq74\" (UID: \"f522807f-660d-4791-95f4-f2d9daa58936\") " pod="openshift-marketplace/community-operators-qnq74" Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.937819 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:31 crc kubenswrapper[4813]: E1206 15:48:31.938113 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:32.438097133 +0000 UTC m=+152.328976699 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:31 crc kubenswrapper[4813]: I1206 15:48:31.979380 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qnq74" Dec 06 15:48:32 crc kubenswrapper[4813]: I1206 15:48:32.000195 4813 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 06 15:48:32 crc kubenswrapper[4813]: I1206 15:48:32.039097 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:32 crc kubenswrapper[4813]: E1206 15:48:32.039427 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:32.539416045 +0000 UTC m=+152.430295621 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:32 crc kubenswrapper[4813]: I1206 15:48:32.140804 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:32 crc kubenswrapper[4813]: E1206 15:48:32.141207 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:32.641192719 +0000 UTC m=+152.532072295 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:32 crc kubenswrapper[4813]: I1206 15:48:32.231387 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wj67k" event={"ID":"0904de59-781e-4c00-86bd-ccebf7a06587","Type":"ContainerStarted","Data":"0ea8aeed2afa32043b42725aefd2fd1456b00c5bddcc37911545e88646e28ea6"} Dec 06 15:48:32 crc kubenswrapper[4813]: I1206 15:48:32.246027 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:32 crc kubenswrapper[4813]: E1206 15:48:32.246397 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:32.746384213 +0000 UTC m=+152.637263789 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:32 crc kubenswrapper[4813]: I1206 15:48:32.271904 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-54xzm"] Dec 06 15:48:32 crc kubenswrapper[4813]: I1206 15:48:32.346751 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:32 crc kubenswrapper[4813]: E1206 15:48:32.348622 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:32.848602209 +0000 UTC m=+152.739481785 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:32 crc kubenswrapper[4813]: I1206 15:48:32.362473 4813 patch_prober.go:28] interesting pod/router-default-5444994796-ncvs7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 15:48:32 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 06 15:48:32 crc kubenswrapper[4813]: [+]process-running ok Dec 06 15:48:32 crc kubenswrapper[4813]: healthz check failed Dec 06 15:48:32 crc kubenswrapper[4813]: I1206 15:48:32.362518 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncvs7" podUID="8ddf8368-8232-4211-bfca-642b1acef6a4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 15:48:32 crc kubenswrapper[4813]: I1206 15:48:32.449972 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:32 crc kubenswrapper[4813]: E1206 15:48:32.450336 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:32.950323431 +0000 UTC m=+152.841203007 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:32 crc kubenswrapper[4813]: I1206 15:48:32.520106 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-wj67k" podStartSLOduration=14.520087088 podStartE2EDuration="14.520087088s" podCreationTimestamp="2025-12-06 15:48:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:32.302536629 +0000 UTC m=+152.193416205" watchObservedRunningTime="2025-12-06 15:48:32.520087088 +0000 UTC m=+152.410966664" Dec 06 15:48:32 crc kubenswrapper[4813]: I1206 15:48:32.521951 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 06 15:48:32 crc kubenswrapper[4813]: W1206 15:48:32.539184 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod977c2e53_21f6_4e10_a539_d57bcb28a6bb.slice/crio-099b940d0a77b7f357c818f6d72d1e658cb45e01200ae605326527082310f5fa WatchSource:0}: Error finding container 099b940d0a77b7f357c818f6d72d1e658cb45e01200ae605326527082310f5fa: Status 404 returned error can't find the container with id 099b940d0a77b7f357c818f6d72d1e658cb45e01200ae605326527082310f5fa Dec 06 15:48:32 crc kubenswrapper[4813]: I1206 15:48:32.550601 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:32 crc kubenswrapper[4813]: E1206 15:48:32.550974 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:33.050957895 +0000 UTC m=+152.941837471 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:32 crc kubenswrapper[4813]: I1206 15:48:32.551083 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:32 crc kubenswrapper[4813]: E1206 15:48:32.551791 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 15:48:33.051782996 +0000 UTC m=+152.942662572 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j79t8" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:32 crc kubenswrapper[4813]: I1206 15:48:32.652219 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:32 crc kubenswrapper[4813]: E1206 15:48:32.652883 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 15:48:33.152865302 +0000 UTC m=+153.043744878 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 15:48:32 crc kubenswrapper[4813]: I1206 15:48:32.738529 4813 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-06T15:48:32.000219447Z","Handler":null,"Name":""} Dec 06 15:48:32 crc kubenswrapper[4813]: I1206 15:48:32.747656 4813 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 06 15:48:32 crc kubenswrapper[4813]: I1206 15:48:32.747693 4813 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 06 15:48:32 crc kubenswrapper[4813]: I1206 15:48:32.755995 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:32 crc kubenswrapper[4813]: I1206 15:48:32.759138 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6rwpr"] Dec 06 15:48:32 crc kubenswrapper[4813]: I1206 15:48:32.795491 4813 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 15:48:32 crc kubenswrapper[4813]: I1206 15:48:32.795536 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:32 crc kubenswrapper[4813]: I1206 15:48:32.911134 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j79t8\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:32 crc kubenswrapper[4813]: I1206 15:48:32.942765 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qnq74"] Dec 06 15:48:32 crc kubenswrapper[4813]: I1206 15:48:32.957674 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.041017 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.087160 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kp6t2"] Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.201371 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.246309 4813 generic.go:334] "Generic (PLEG): container finished" podID="e2fac5b3-c60b-4c96-8841-ae53a24aaed1" containerID="3239e6c2ed4ba462e7826d0ff5613cabec75c1a9849a92ccc7f70500ea73c40c" exitCode=0 Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.246371 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-54xzm" event={"ID":"e2fac5b3-c60b-4c96-8841-ae53a24aaed1","Type":"ContainerDied","Data":"3239e6c2ed4ba462e7826d0ff5613cabec75c1a9849a92ccc7f70500ea73c40c"} Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.246396 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-54xzm" event={"ID":"e2fac5b3-c60b-4c96-8841-ae53a24aaed1","Type":"ContainerStarted","Data":"bff1c33004538532a4b7d28d51ac7266dd442a1933568cb36ecb5430816e9a09"} Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.249500 4813 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.253586 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kp6t2" event={"ID":"e982bb16-5f82-418b-8c03-8c73d22e010b","Type":"ContainerStarted","Data":"dc06103c0a720defb5a411399c5900332373d97efcbe58424f0879cd6c72645b"} Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.255138 4813 generic.go:334] "Generic (PLEG): container finished" podID="f522807f-660d-4791-95f4-f2d9daa58936" containerID="ad14489bc0eea912c65421f84a92ce0d9ae9028f677db47ed06e5142e4929884" exitCode=0 Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.255177 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qnq74" event={"ID":"f522807f-660d-4791-95f4-f2d9daa58936","Type":"ContainerDied","Data":"ad14489bc0eea912c65421f84a92ce0d9ae9028f677db47ed06e5142e4929884"} Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.255191 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qnq74" event={"ID":"f522807f-660d-4791-95f4-f2d9daa58936","Type":"ContainerStarted","Data":"c1fe3e4f52a8a698a90dfd2bff093f40e85bc638cf468d8f9f9e8f43f475a0f2"} Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.259645 4813 generic.go:334] "Generic (PLEG): container finished" podID="76073738-387a-4968-a483-8880e866d1e4" containerID="bc96f70a14b2fbc8f38437bcc68ace07662e44f0bf69a3bb3a0824987b67e5a1" exitCode=0 Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.259714 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rwpr" event={"ID":"76073738-387a-4968-a483-8880e866d1e4","Type":"ContainerDied","Data":"bc96f70a14b2fbc8f38437bcc68ace07662e44f0bf69a3bb3a0824987b67e5a1"} Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.259740 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rwpr" event={"ID":"76073738-387a-4968-a483-8880e866d1e4","Type":"ContainerStarted","Data":"c3981bff3b38eea82b4e02f97eb87c8a7b9cf407ba1889805a8b471eaf924e70"} Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.263162 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"977c2e53-21f6-4e10-a539-d57bcb28a6bb","Type":"ContainerStarted","Data":"c959ba1e91672a3fa4c7e5720910c18ddc1899144b63b31fe5511b59c86b7afb"} Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.263188 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"977c2e53-21f6-4e10-a539-d57bcb28a6bb","Type":"ContainerStarted","Data":"099b940d0a77b7f357c818f6d72d1e658cb45e01200ae605326527082310f5fa"} Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.344565 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.34454686 podStartE2EDuration="3.34454686s" podCreationTimestamp="2025-12-06 15:48:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:33.337596426 +0000 UTC m=+153.228476012" watchObservedRunningTime="2025-12-06 15:48:33.34454686 +0000 UTC m=+153.235426436" Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.349967 4813 patch_prober.go:28] interesting pod/router-default-5444994796-ncvs7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 15:48:33 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 06 15:48:33 crc kubenswrapper[4813]: [+]process-running ok Dec 06 15:48:33 crc kubenswrapper[4813]: healthz check failed Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.350027 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncvs7" podUID="8ddf8368-8232-4211-bfca-642b1acef6a4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.541912 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-j79t8"] Dec 06 15:48:33 crc kubenswrapper[4813]: W1206 15:48:33.550749 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd2857e44_0b98_4f13_baca_3330474d8b44.slice/crio-4c61e905c9692c4d5c5e38a288853af45e48b19329159f50a06490cabdaf0f01 WatchSource:0}: Error finding container 4c61e905c9692c4d5c5e38a288853af45e48b19329159f50a06490cabdaf0f01: Status 404 returned error can't find the container with id 4c61e905c9692c4d5c5e38a288853af45e48b19329159f50a06490cabdaf0f01 Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.573060 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vn7fl"] Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.629580 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vn7fl"] Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.629791 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vn7fl" Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.632128 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.767288 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5zp6r"] Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.769046 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5zp6r" Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.775284 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da4de6a4-9efc-425c-8712-e66919babdf2-utilities\") pod \"redhat-marketplace-vn7fl\" (UID: \"da4de6a4-9efc-425c-8712-e66919babdf2\") " pod="openshift-marketplace/redhat-marketplace-vn7fl" Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.775350 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xft4q\" (UniqueName: \"kubernetes.io/projected/da4de6a4-9efc-425c-8712-e66919babdf2-kube-api-access-xft4q\") pod \"redhat-marketplace-vn7fl\" (UID: \"da4de6a4-9efc-425c-8712-e66919babdf2\") " pod="openshift-marketplace/redhat-marketplace-vn7fl" Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.775420 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da4de6a4-9efc-425c-8712-e66919babdf2-catalog-content\") pod \"redhat-marketplace-vn7fl\" (UID: \"da4de6a4-9efc-425c-8712-e66919babdf2\") " pod="openshift-marketplace/redhat-marketplace-vn7fl" Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.777220 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.781755 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5zp6r"] Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.876706 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kv5hk\" (UniqueName: \"kubernetes.io/projected/6a553ef5-b76a-46cd-9e49-30bf4ca86e29-kube-api-access-kv5hk\") pod \"redhat-operators-5zp6r\" (UID: \"6a553ef5-b76a-46cd-9e49-30bf4ca86e29\") " pod="openshift-marketplace/redhat-operators-5zp6r" Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.877042 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da4de6a4-9efc-425c-8712-e66919babdf2-utilities\") pod \"redhat-marketplace-vn7fl\" (UID: \"da4de6a4-9efc-425c-8712-e66919babdf2\") " pod="openshift-marketplace/redhat-marketplace-vn7fl" Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.877074 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xft4q\" (UniqueName: \"kubernetes.io/projected/da4de6a4-9efc-425c-8712-e66919babdf2-kube-api-access-xft4q\") pod \"redhat-marketplace-vn7fl\" (UID: \"da4de6a4-9efc-425c-8712-e66919babdf2\") " pod="openshift-marketplace/redhat-marketplace-vn7fl" Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.877138 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a553ef5-b76a-46cd-9e49-30bf4ca86e29-utilities\") pod \"redhat-operators-5zp6r\" (UID: \"6a553ef5-b76a-46cd-9e49-30bf4ca86e29\") " pod="openshift-marketplace/redhat-operators-5zp6r" Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.877194 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a553ef5-b76a-46cd-9e49-30bf4ca86e29-catalog-content\") pod \"redhat-operators-5zp6r\" (UID: \"6a553ef5-b76a-46cd-9e49-30bf4ca86e29\") " pod="openshift-marketplace/redhat-operators-5zp6r" Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.877220 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da4de6a4-9efc-425c-8712-e66919babdf2-catalog-content\") pod \"redhat-marketplace-vn7fl\" (UID: \"da4de6a4-9efc-425c-8712-e66919babdf2\") " pod="openshift-marketplace/redhat-marketplace-vn7fl" Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.877685 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da4de6a4-9efc-425c-8712-e66919babdf2-catalog-content\") pod \"redhat-marketplace-vn7fl\" (UID: \"da4de6a4-9efc-425c-8712-e66919babdf2\") " pod="openshift-marketplace/redhat-marketplace-vn7fl" Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.877997 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da4de6a4-9efc-425c-8712-e66919babdf2-utilities\") pod \"redhat-marketplace-vn7fl\" (UID: \"da4de6a4-9efc-425c-8712-e66919babdf2\") " pod="openshift-marketplace/redhat-marketplace-vn7fl" Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.917185 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xft4q\" (UniqueName: \"kubernetes.io/projected/da4de6a4-9efc-425c-8712-e66919babdf2-kube-api-access-xft4q\") pod \"redhat-marketplace-vn7fl\" (UID: \"da4de6a4-9efc-425c-8712-e66919babdf2\") " pod="openshift-marketplace/redhat-marketplace-vn7fl" Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.978619 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kv5hk\" (UniqueName: \"kubernetes.io/projected/6a553ef5-b76a-46cd-9e49-30bf4ca86e29-kube-api-access-kv5hk\") pod \"redhat-operators-5zp6r\" (UID: \"6a553ef5-b76a-46cd-9e49-30bf4ca86e29\") " pod="openshift-marketplace/redhat-operators-5zp6r" Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.978721 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a553ef5-b76a-46cd-9e49-30bf4ca86e29-utilities\") pod \"redhat-operators-5zp6r\" (UID: \"6a553ef5-b76a-46cd-9e49-30bf4ca86e29\") " pod="openshift-marketplace/redhat-operators-5zp6r" Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.978757 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a553ef5-b76a-46cd-9e49-30bf4ca86e29-catalog-content\") pod \"redhat-operators-5zp6r\" (UID: \"6a553ef5-b76a-46cd-9e49-30bf4ca86e29\") " pod="openshift-marketplace/redhat-operators-5zp6r" Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.979197 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a553ef5-b76a-46cd-9e49-30bf4ca86e29-catalog-content\") pod \"redhat-operators-5zp6r\" (UID: \"6a553ef5-b76a-46cd-9e49-30bf4ca86e29\") " pod="openshift-marketplace/redhat-operators-5zp6r" Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.979173 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a553ef5-b76a-46cd-9e49-30bf4ca86e29-utilities\") pod \"redhat-operators-5zp6r\" (UID: \"6a553ef5-b76a-46cd-9e49-30bf4ca86e29\") " pod="openshift-marketplace/redhat-operators-5zp6r" Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.987541 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pvphv"] Dec 06 15:48:33 crc kubenswrapper[4813]: I1206 15:48:33.988698 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pvphv" Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.034610 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vn7fl" Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.035735 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kv5hk\" (UniqueName: \"kubernetes.io/projected/6a553ef5-b76a-46cd-9e49-30bf4ca86e29-kube-api-access-kv5hk\") pod \"redhat-operators-5zp6r\" (UID: \"6a553ef5-b76a-46cd-9e49-30bf4ca86e29\") " pod="openshift-marketplace/redhat-operators-5zp6r" Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.069127 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pvphv"] Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.111772 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e2017ee-a77f-47b3-ae48-058c62ce30d6-utilities\") pod \"redhat-marketplace-pvphv\" (UID: \"7e2017ee-a77f-47b3-ae48-058c62ce30d6\") " pod="openshift-marketplace/redhat-marketplace-pvphv" Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.111848 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-445mw\" (UniqueName: \"kubernetes.io/projected/7e2017ee-a77f-47b3-ae48-058c62ce30d6-kube-api-access-445mw\") pod \"redhat-marketplace-pvphv\" (UID: \"7e2017ee-a77f-47b3-ae48-058c62ce30d6\") " pod="openshift-marketplace/redhat-marketplace-pvphv" Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.111869 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e2017ee-a77f-47b3-ae48-058c62ce30d6-catalog-content\") pod \"redhat-marketplace-pvphv\" (UID: \"7e2017ee-a77f-47b3-ae48-058c62ce30d6\") " pod="openshift-marketplace/redhat-marketplace-pvphv" Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.112092 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5zp6r" Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.164415 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7d6jw"] Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.173462 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7d6jw" Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.213213 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e2017ee-a77f-47b3-ae48-058c62ce30d6-utilities\") pod \"redhat-marketplace-pvphv\" (UID: \"7e2017ee-a77f-47b3-ae48-058c62ce30d6\") " pod="openshift-marketplace/redhat-marketplace-pvphv" Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.213650 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e2017ee-a77f-47b3-ae48-058c62ce30d6-utilities\") pod \"redhat-marketplace-pvphv\" (UID: \"7e2017ee-a77f-47b3-ae48-058c62ce30d6\") " pod="openshift-marketplace/redhat-marketplace-pvphv" Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.213736 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-445mw\" (UniqueName: \"kubernetes.io/projected/7e2017ee-a77f-47b3-ae48-058c62ce30d6-kube-api-access-445mw\") pod \"redhat-marketplace-pvphv\" (UID: \"7e2017ee-a77f-47b3-ae48-058c62ce30d6\") " pod="openshift-marketplace/redhat-marketplace-pvphv" Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.213761 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e2017ee-a77f-47b3-ae48-058c62ce30d6-catalog-content\") pod \"redhat-marketplace-pvphv\" (UID: \"7e2017ee-a77f-47b3-ae48-058c62ce30d6\") " pod="openshift-marketplace/redhat-marketplace-pvphv" Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.214293 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e2017ee-a77f-47b3-ae48-058c62ce30d6-catalog-content\") pod \"redhat-marketplace-pvphv\" (UID: \"7e2017ee-a77f-47b3-ae48-058c62ce30d6\") " pod="openshift-marketplace/redhat-marketplace-pvphv" Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.227679 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7d6jw"] Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.242876 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-445mw\" (UniqueName: \"kubernetes.io/projected/7e2017ee-a77f-47b3-ae48-058c62ce30d6-kube-api-access-445mw\") pod \"redhat-marketplace-pvphv\" (UID: \"7e2017ee-a77f-47b3-ae48-058c62ce30d6\") " pod="openshift-marketplace/redhat-marketplace-pvphv" Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.306664 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pvphv" Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.315923 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/952e1f68-97e3-4191-9dde-43d67e83a1e0-catalog-content\") pod \"redhat-operators-7d6jw\" (UID: \"952e1f68-97e3-4191-9dde-43d67e83a1e0\") " pod="openshift-marketplace/redhat-operators-7d6jw" Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.316017 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqpc5\" (UniqueName: \"kubernetes.io/projected/952e1f68-97e3-4191-9dde-43d67e83a1e0-kube-api-access-pqpc5\") pod \"redhat-operators-7d6jw\" (UID: \"952e1f68-97e3-4191-9dde-43d67e83a1e0\") " pod="openshift-marketplace/redhat-operators-7d6jw" Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.316066 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/952e1f68-97e3-4191-9dde-43d67e83a1e0-utilities\") pod \"redhat-operators-7d6jw\" (UID: \"952e1f68-97e3-4191-9dde-43d67e83a1e0\") " pod="openshift-marketplace/redhat-operators-7d6jw" Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.348298 4813 patch_prober.go:28] interesting pod/router-default-5444994796-ncvs7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 15:48:34 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 06 15:48:34 crc kubenswrapper[4813]: [+]process-running ok Dec 06 15:48:34 crc kubenswrapper[4813]: healthz check failed Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.348348 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncvs7" podUID="8ddf8368-8232-4211-bfca-642b1acef6a4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.417738 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/952e1f68-97e3-4191-9dde-43d67e83a1e0-catalog-content\") pod \"redhat-operators-7d6jw\" (UID: \"952e1f68-97e3-4191-9dde-43d67e83a1e0\") " pod="openshift-marketplace/redhat-operators-7d6jw" Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.417803 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqpc5\" (UniqueName: \"kubernetes.io/projected/952e1f68-97e3-4191-9dde-43d67e83a1e0-kube-api-access-pqpc5\") pod \"redhat-operators-7d6jw\" (UID: \"952e1f68-97e3-4191-9dde-43d67e83a1e0\") " pod="openshift-marketplace/redhat-operators-7d6jw" Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.417839 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/952e1f68-97e3-4191-9dde-43d67e83a1e0-utilities\") pod \"redhat-operators-7d6jw\" (UID: \"952e1f68-97e3-4191-9dde-43d67e83a1e0\") " pod="openshift-marketplace/redhat-operators-7d6jw" Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.418357 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/952e1f68-97e3-4191-9dde-43d67e83a1e0-utilities\") pod \"redhat-operators-7d6jw\" (UID: \"952e1f68-97e3-4191-9dde-43d67e83a1e0\") " pod="openshift-marketplace/redhat-operators-7d6jw" Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.418431 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/952e1f68-97e3-4191-9dde-43d67e83a1e0-catalog-content\") pod \"redhat-operators-7d6jw\" (UID: \"952e1f68-97e3-4191-9dde-43d67e83a1e0\") " pod="openshift-marketplace/redhat-operators-7d6jw" Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.457101 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqpc5\" (UniqueName: \"kubernetes.io/projected/952e1f68-97e3-4191-9dde-43d67e83a1e0-kube-api-access-pqpc5\") pod \"redhat-operators-7d6jw\" (UID: \"952e1f68-97e3-4191-9dde-43d67e83a1e0\") " pod="openshift-marketplace/redhat-operators-7d6jw" Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.466902 4813 generic.go:334] "Generic (PLEG): container finished" podID="977c2e53-21f6-4e10-a539-d57bcb28a6bb" containerID="c959ba1e91672a3fa4c7e5720910c18ddc1899144b63b31fe5511b59c86b7afb" exitCode=0 Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.467164 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"977c2e53-21f6-4e10-a539-d57bcb28a6bb","Type":"ContainerDied","Data":"c959ba1e91672a3fa4c7e5720910c18ddc1899144b63b31fe5511b59c86b7afb"} Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.486666 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" event={"ID":"d2857e44-0b98-4f13-baca-3330474d8b44","Type":"ContainerStarted","Data":"776db7e851680747dc3871ee8baa6b50b31a045938289c74570f864d7d66dbf8"} Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.505522 4813 generic.go:334] "Generic (PLEG): container finished" podID="e982bb16-5f82-418b-8c03-8c73d22e010b" containerID="a76753c7cb31be3afa135159cb50365b7cb5e7f2c369b2f81b2b7e4a7b2deb76" exitCode=0 Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.519464 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7d6jw" Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.535788 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.536457 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.536493 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" event={"ID":"d2857e44-0b98-4f13-baca-3330474d8b44","Type":"ContainerStarted","Data":"4c61e905c9692c4d5c5e38a288853af45e48b19329159f50a06490cabdaf0f01"} Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.536507 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kp6t2" event={"ID":"e982bb16-5f82-418b-8c03-8c73d22e010b","Type":"ContainerDied","Data":"a76753c7cb31be3afa135159cb50365b7cb5e7f2c369b2f81b2b7e4a7b2deb76"} Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.538432 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" podStartSLOduration=134.538416039 podStartE2EDuration="2m14.538416039s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:34.534742742 +0000 UTC m=+154.425622318" watchObservedRunningTime="2025-12-06 15:48:34.538416039 +0000 UTC m=+154.429295615" Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.582645 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vn7fl"] Dec 06 15:48:34 crc kubenswrapper[4813]: I1206 15:48:34.708583 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5zp6r"] Dec 06 15:48:35 crc kubenswrapper[4813]: I1206 15:48:35.059010 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7d6jw"] Dec 06 15:48:35 crc kubenswrapper[4813]: W1206 15:48:35.083281 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod952e1f68_97e3_4191_9dde_43d67e83a1e0.slice/crio-a22f4bb1238a969321aed571002743042e17a097213d53ae9ab79abd34b9da6c WatchSource:0}: Error finding container a22f4bb1238a969321aed571002743042e17a097213d53ae9ab79abd34b9da6c: Status 404 returned error can't find the container with id a22f4bb1238a969321aed571002743042e17a097213d53ae9ab79abd34b9da6c Dec 06 15:48:35 crc kubenswrapper[4813]: I1206 15:48:35.182169 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pvphv"] Dec 06 15:48:35 crc kubenswrapper[4813]: I1206 15:48:35.216547 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 06 15:48:35 crc kubenswrapper[4813]: I1206 15:48:35.217180 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 15:48:35 crc kubenswrapper[4813]: I1206 15:48:35.218726 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 06 15:48:35 crc kubenswrapper[4813]: I1206 15:48:35.218799 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 06 15:48:35 crc kubenswrapper[4813]: W1206 15:48:35.222547 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e2017ee_a77f_47b3_ae48_058c62ce30d6.slice/crio-9dc3b473db2f7f0e842d40c093a40b6cc5f7da306114acbfe2632a40859abbe0 WatchSource:0}: Error finding container 9dc3b473db2f7f0e842d40c093a40b6cc5f7da306114acbfe2632a40859abbe0: Status 404 returned error can't find the container with id 9dc3b473db2f7f0e842d40c093a40b6cc5f7da306114acbfe2632a40859abbe0 Dec 06 15:48:35 crc kubenswrapper[4813]: I1206 15:48:35.232275 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 06 15:48:35 crc kubenswrapper[4813]: I1206 15:48:35.341272 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3ac08880-2cdb-4dc7-a62a-d68e23e7919b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"3ac08880-2cdb-4dc7-a62a-d68e23e7919b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 15:48:35 crc kubenswrapper[4813]: I1206 15:48:35.341776 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3ac08880-2cdb-4dc7-a62a-d68e23e7919b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"3ac08880-2cdb-4dc7-a62a-d68e23e7919b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 15:48:35 crc kubenswrapper[4813]: I1206 15:48:35.349426 4813 patch_prober.go:28] interesting pod/router-default-5444994796-ncvs7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 15:48:35 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 06 15:48:35 crc kubenswrapper[4813]: [+]process-running ok Dec 06 15:48:35 crc kubenswrapper[4813]: healthz check failed Dec 06 15:48:35 crc kubenswrapper[4813]: I1206 15:48:35.349475 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncvs7" podUID="8ddf8368-8232-4211-bfca-642b1acef6a4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 15:48:35 crc kubenswrapper[4813]: I1206 15:48:35.447809 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3ac08880-2cdb-4dc7-a62a-d68e23e7919b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"3ac08880-2cdb-4dc7-a62a-d68e23e7919b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 15:48:35 crc kubenswrapper[4813]: I1206 15:48:35.447866 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3ac08880-2cdb-4dc7-a62a-d68e23e7919b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"3ac08880-2cdb-4dc7-a62a-d68e23e7919b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 15:48:35 crc kubenswrapper[4813]: I1206 15:48:35.448335 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3ac08880-2cdb-4dc7-a62a-d68e23e7919b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"3ac08880-2cdb-4dc7-a62a-d68e23e7919b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 15:48:35 crc kubenswrapper[4813]: I1206 15:48:35.479360 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3ac08880-2cdb-4dc7-a62a-d68e23e7919b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"3ac08880-2cdb-4dc7-a62a-d68e23e7919b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 15:48:35 crc kubenswrapper[4813]: I1206 15:48:35.497694 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:35 crc kubenswrapper[4813]: I1206 15:48:35.501835 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-8vzrm" Dec 06 15:48:35 crc kubenswrapper[4813]: I1206 15:48:35.513340 4813 generic.go:334] "Generic (PLEG): container finished" podID="952e1f68-97e3-4191-9dde-43d67e83a1e0" containerID="f92e84ded684cbb4d8495f28050e654be29b0c5ea9192efc0803cba604d48054" exitCode=0 Dec 06 15:48:35 crc kubenswrapper[4813]: I1206 15:48:35.513400 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7d6jw" event={"ID":"952e1f68-97e3-4191-9dde-43d67e83a1e0","Type":"ContainerDied","Data":"f92e84ded684cbb4d8495f28050e654be29b0c5ea9192efc0803cba604d48054"} Dec 06 15:48:35 crc kubenswrapper[4813]: I1206 15:48:35.515446 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7d6jw" event={"ID":"952e1f68-97e3-4191-9dde-43d67e83a1e0","Type":"ContainerStarted","Data":"a22f4bb1238a969321aed571002743042e17a097213d53ae9ab79abd34b9da6c"} Dec 06 15:48:35 crc kubenswrapper[4813]: I1206 15:48:35.536682 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pvphv" event={"ID":"7e2017ee-a77f-47b3-ae48-058c62ce30d6","Type":"ContainerStarted","Data":"9dc3b473db2f7f0e842d40c093a40b6cc5f7da306114acbfe2632a40859abbe0"} Dec 06 15:48:35 crc kubenswrapper[4813]: I1206 15:48:35.539708 4813 generic.go:334] "Generic (PLEG): container finished" podID="da4de6a4-9efc-425c-8712-e66919babdf2" containerID="9ddb470f673e893e2ff4c07442e80dc9b2c05b178f3fe952c272878d080a1bdb" exitCode=0 Dec 06 15:48:35 crc kubenswrapper[4813]: I1206 15:48:35.539763 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vn7fl" event={"ID":"da4de6a4-9efc-425c-8712-e66919babdf2","Type":"ContainerDied","Data":"9ddb470f673e893e2ff4c07442e80dc9b2c05b178f3fe952c272878d080a1bdb"} Dec 06 15:48:35 crc kubenswrapper[4813]: I1206 15:48:35.539788 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vn7fl" event={"ID":"da4de6a4-9efc-425c-8712-e66919babdf2","Type":"ContainerStarted","Data":"6d7c0708da894695c1752c4c837caba78ff36fcd0029b9e127a7c99521af7cce"} Dec 06 15:48:35 crc kubenswrapper[4813]: I1206 15:48:35.550820 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 15:48:35 crc kubenswrapper[4813]: I1206 15:48:35.552974 4813 generic.go:334] "Generic (PLEG): container finished" podID="6a553ef5-b76a-46cd-9e49-30bf4ca86e29" containerID="88d68d018baeb0bf77b1b600c7fd176ca919cf44a090702ca393ca3c59fec7bd" exitCode=0 Dec 06 15:48:35 crc kubenswrapper[4813]: I1206 15:48:35.553331 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5zp6r" event={"ID":"6a553ef5-b76a-46cd-9e49-30bf4ca86e29","Type":"ContainerDied","Data":"88d68d018baeb0bf77b1b600c7fd176ca919cf44a090702ca393ca3c59fec7bd"} Dec 06 15:48:35 crc kubenswrapper[4813]: I1206 15:48:35.553384 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5zp6r" event={"ID":"6a553ef5-b76a-46cd-9e49-30bf4ca86e29","Type":"ContainerStarted","Data":"9b87ae2c54a10257d7fd192ae57c948eaeadaeba991d0c9e46506b682085ece9"} Dec 06 15:48:36 crc kubenswrapper[4813]: I1206 15:48:36.286410 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 15:48:36 crc kubenswrapper[4813]: I1206 15:48:36.345045 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 06 15:48:36 crc kubenswrapper[4813]: I1206 15:48:36.347614 4813 patch_prober.go:28] interesting pod/router-default-5444994796-ncvs7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 15:48:36 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 06 15:48:36 crc kubenswrapper[4813]: [+]process-running ok Dec 06 15:48:36 crc kubenswrapper[4813]: healthz check failed Dec 06 15:48:36 crc kubenswrapper[4813]: I1206 15:48:36.347653 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncvs7" podUID="8ddf8368-8232-4211-bfca-642b1acef6a4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 15:48:36 crc kubenswrapper[4813]: W1206 15:48:36.374911 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod3ac08880_2cdb_4dc7_a62a_d68e23e7919b.slice/crio-54cfc28490e555375c706660e61a1dc12dca009bbc9b5a12b18cf8ff5d7a998e WatchSource:0}: Error finding container 54cfc28490e555375c706660e61a1dc12dca009bbc9b5a12b18cf8ff5d7a998e: Status 404 returned error can't find the container with id 54cfc28490e555375c706660e61a1dc12dca009bbc9b5a12b18cf8ff5d7a998e Dec 06 15:48:36 crc kubenswrapper[4813]: I1206 15:48:36.383591 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/977c2e53-21f6-4e10-a539-d57bcb28a6bb-kubelet-dir\") pod \"977c2e53-21f6-4e10-a539-d57bcb28a6bb\" (UID: \"977c2e53-21f6-4e10-a539-d57bcb28a6bb\") " Dec 06 15:48:36 crc kubenswrapper[4813]: I1206 15:48:36.383723 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/977c2e53-21f6-4e10-a539-d57bcb28a6bb-kube-api-access\") pod \"977c2e53-21f6-4e10-a539-d57bcb28a6bb\" (UID: \"977c2e53-21f6-4e10-a539-d57bcb28a6bb\") " Dec 06 15:48:36 crc kubenswrapper[4813]: I1206 15:48:36.383918 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/977c2e53-21f6-4e10-a539-d57bcb28a6bb-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "977c2e53-21f6-4e10-a539-d57bcb28a6bb" (UID: "977c2e53-21f6-4e10-a539-d57bcb28a6bb"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 15:48:36 crc kubenswrapper[4813]: I1206 15:48:36.384088 4813 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/977c2e53-21f6-4e10-a539-d57bcb28a6bb-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 15:48:36 crc kubenswrapper[4813]: I1206 15:48:36.403848 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/977c2e53-21f6-4e10-a539-d57bcb28a6bb-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "977c2e53-21f6-4e10-a539-d57bcb28a6bb" (UID: "977c2e53-21f6-4e10-a539-d57bcb28a6bb"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:48:36 crc kubenswrapper[4813]: I1206 15:48:36.486218 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/977c2e53-21f6-4e10-a539-d57bcb28a6bb-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 15:48:36 crc kubenswrapper[4813]: I1206 15:48:36.570627 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"977c2e53-21f6-4e10-a539-d57bcb28a6bb","Type":"ContainerDied","Data":"099b940d0a77b7f357c818f6d72d1e658cb45e01200ae605326527082310f5fa"} Dec 06 15:48:36 crc kubenswrapper[4813]: I1206 15:48:36.570664 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="099b940d0a77b7f357c818f6d72d1e658cb45e01200ae605326527082310f5fa" Dec 06 15:48:36 crc kubenswrapper[4813]: I1206 15:48:36.570720 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 15:48:36 crc kubenswrapper[4813]: I1206 15:48:36.577510 4813 generic.go:334] "Generic (PLEG): container finished" podID="7e2017ee-a77f-47b3-ae48-058c62ce30d6" containerID="ff2d29a7f28dd730b01271bf394ad2bbb513ece6aa593c67a8793e32f5e407bd" exitCode=0 Dec 06 15:48:36 crc kubenswrapper[4813]: I1206 15:48:36.577559 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pvphv" event={"ID":"7e2017ee-a77f-47b3-ae48-058c62ce30d6","Type":"ContainerDied","Data":"ff2d29a7f28dd730b01271bf394ad2bbb513ece6aa593c67a8793e32f5e407bd"} Dec 06 15:48:36 crc kubenswrapper[4813]: I1206 15:48:36.579723 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"3ac08880-2cdb-4dc7-a62a-d68e23e7919b","Type":"ContainerStarted","Data":"54cfc28490e555375c706660e61a1dc12dca009bbc9b5a12b18cf8ff5d7a998e"} Dec 06 15:48:37 crc kubenswrapper[4813]: I1206 15:48:37.352993 4813 patch_prober.go:28] interesting pod/router-default-5444994796-ncvs7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 15:48:37 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 06 15:48:37 crc kubenswrapper[4813]: [+]process-running ok Dec 06 15:48:37 crc kubenswrapper[4813]: healthz check failed Dec 06 15:48:37 crc kubenswrapper[4813]: I1206 15:48:37.353282 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncvs7" podUID="8ddf8368-8232-4211-bfca-642b1acef6a4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 15:48:37 crc kubenswrapper[4813]: I1206 15:48:37.616911 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"3ac08880-2cdb-4dc7-a62a-d68e23e7919b","Type":"ContainerStarted","Data":"595a2f567d0845c0515973e4406ccee83167139128edb5ea0286e9dcfa1a5732"} Dec 06 15:48:37 crc kubenswrapper[4813]: I1206 15:48:37.618112 4813 generic.go:334] "Generic (PLEG): container finished" podID="8810f12c-6dbc-4bf9-b27e-29ebc5986955" containerID="2fba0692193bb5fa545d890a31ff21e880014c4c26c01164df7322b66f8ec130" exitCode=0 Dec 06 15:48:37 crc kubenswrapper[4813]: I1206 15:48:37.618136 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417265-hcswl" event={"ID":"8810f12c-6dbc-4bf9-b27e-29ebc5986955","Type":"ContainerDied","Data":"2fba0692193bb5fa545d890a31ff21e880014c4c26c01164df7322b66f8ec130"} Dec 06 15:48:38 crc kubenswrapper[4813]: I1206 15:48:38.366233 4813 patch_prober.go:28] interesting pod/router-default-5444994796-ncvs7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 15:48:38 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 06 15:48:38 crc kubenswrapper[4813]: [+]process-running ok Dec 06 15:48:38 crc kubenswrapper[4813]: healthz check failed Dec 06 15:48:38 crc kubenswrapper[4813]: I1206 15:48:38.366657 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncvs7" podUID="8ddf8368-8232-4211-bfca-642b1acef6a4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 15:48:39 crc kubenswrapper[4813]: I1206 15:48:39.323194 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417265-hcswl" Dec 06 15:48:39 crc kubenswrapper[4813]: I1206 15:48:39.363568 4813 patch_prober.go:28] interesting pod/router-default-5444994796-ncvs7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 15:48:39 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 06 15:48:39 crc kubenswrapper[4813]: [+]process-running ok Dec 06 15:48:39 crc kubenswrapper[4813]: healthz check failed Dec 06 15:48:39 crc kubenswrapper[4813]: I1206 15:48:39.363645 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncvs7" podUID="8ddf8368-8232-4211-bfca-642b1acef6a4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 15:48:39 crc kubenswrapper[4813]: I1206 15:48:39.388694 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=4.388673149 podStartE2EDuration="4.388673149s" podCreationTimestamp="2025-12-06 15:48:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:38.644150522 +0000 UTC m=+158.535030088" watchObservedRunningTime="2025-12-06 15:48:39.388673149 +0000 UTC m=+159.279552725" Dec 06 15:48:39 crc kubenswrapper[4813]: I1206 15:48:39.438930 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-tm2zm" Dec 06 15:48:39 crc kubenswrapper[4813]: I1206 15:48:39.490989 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8810f12c-6dbc-4bf9-b27e-29ebc5986955-secret-volume\") pod \"8810f12c-6dbc-4bf9-b27e-29ebc5986955\" (UID: \"8810f12c-6dbc-4bf9-b27e-29ebc5986955\") " Dec 06 15:48:39 crc kubenswrapper[4813]: I1206 15:48:39.491074 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6n976\" (UniqueName: \"kubernetes.io/projected/8810f12c-6dbc-4bf9-b27e-29ebc5986955-kube-api-access-6n976\") pod \"8810f12c-6dbc-4bf9-b27e-29ebc5986955\" (UID: \"8810f12c-6dbc-4bf9-b27e-29ebc5986955\") " Dec 06 15:48:39 crc kubenswrapper[4813]: I1206 15:48:39.491106 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8810f12c-6dbc-4bf9-b27e-29ebc5986955-config-volume\") pod \"8810f12c-6dbc-4bf9-b27e-29ebc5986955\" (UID: \"8810f12c-6dbc-4bf9-b27e-29ebc5986955\") " Dec 06 15:48:39 crc kubenswrapper[4813]: I1206 15:48:39.492058 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8810f12c-6dbc-4bf9-b27e-29ebc5986955-config-volume" (OuterVolumeSpecName: "config-volume") pod "8810f12c-6dbc-4bf9-b27e-29ebc5986955" (UID: "8810f12c-6dbc-4bf9-b27e-29ebc5986955"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:48:39 crc kubenswrapper[4813]: I1206 15:48:39.517907 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8810f12c-6dbc-4bf9-b27e-29ebc5986955-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8810f12c-6dbc-4bf9-b27e-29ebc5986955" (UID: "8810f12c-6dbc-4bf9-b27e-29ebc5986955"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:48:39 crc kubenswrapper[4813]: I1206 15:48:39.553228 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8810f12c-6dbc-4bf9-b27e-29ebc5986955-kube-api-access-6n976" (OuterVolumeSpecName: "kube-api-access-6n976") pod "8810f12c-6dbc-4bf9-b27e-29ebc5986955" (UID: "8810f12c-6dbc-4bf9-b27e-29ebc5986955"). InnerVolumeSpecName "kube-api-access-6n976". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:48:39 crc kubenswrapper[4813]: I1206 15:48:39.592612 4813 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8810f12c-6dbc-4bf9-b27e-29ebc5986955-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 15:48:39 crc kubenswrapper[4813]: I1206 15:48:39.592641 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6n976\" (UniqueName: \"kubernetes.io/projected/8810f12c-6dbc-4bf9-b27e-29ebc5986955-kube-api-access-6n976\") on node \"crc\" DevicePath \"\"" Dec 06 15:48:39 crc kubenswrapper[4813]: I1206 15:48:39.592650 4813 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8810f12c-6dbc-4bf9-b27e-29ebc5986955-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 15:48:39 crc kubenswrapper[4813]: I1206 15:48:39.716886 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417265-hcswl" Dec 06 15:48:39 crc kubenswrapper[4813]: I1206 15:48:39.716875 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417265-hcswl" event={"ID":"8810f12c-6dbc-4bf9-b27e-29ebc5986955","Type":"ContainerDied","Data":"7c38aa4dffc07b1b3e5abbe869db360dcb5fa4bace0370a0c752b04a317dce13"} Dec 06 15:48:39 crc kubenswrapper[4813]: I1206 15:48:39.717700 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c38aa4dffc07b1b3e5abbe869db360dcb5fa4bace0370a0c752b04a317dce13" Dec 06 15:48:39 crc kubenswrapper[4813]: I1206 15:48:39.746346 4813 generic.go:334] "Generic (PLEG): container finished" podID="3ac08880-2cdb-4dc7-a62a-d68e23e7919b" containerID="595a2f567d0845c0515973e4406ccee83167139128edb5ea0286e9dcfa1a5732" exitCode=0 Dec 06 15:48:39 crc kubenswrapper[4813]: I1206 15:48:39.746668 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"3ac08880-2cdb-4dc7-a62a-d68e23e7919b","Type":"ContainerDied","Data":"595a2f567d0845c0515973e4406ccee83167139128edb5ea0286e9dcfa1a5732"} Dec 06 15:48:40 crc kubenswrapper[4813]: I1206 15:48:40.361426 4813 patch_prober.go:28] interesting pod/router-default-5444994796-ncvs7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 15:48:40 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 06 15:48:40 crc kubenswrapper[4813]: [+]process-running ok Dec 06 15:48:40 crc kubenswrapper[4813]: healthz check failed Dec 06 15:48:40 crc kubenswrapper[4813]: I1206 15:48:40.361570 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncvs7" podUID="8ddf8368-8232-4211-bfca-642b1acef6a4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 15:48:40 crc kubenswrapper[4813]: I1206 15:48:40.511317 4813 patch_prober.go:28] interesting pod/console-f9d7485db-gb6cc container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Dec 06 15:48:40 crc kubenswrapper[4813]: I1206 15:48:40.511374 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-gb6cc" podUID="36f25c5a-3e57-4367-b306-db9661e4f7c9" containerName="console" probeResult="failure" output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" Dec 06 15:48:41 crc kubenswrapper[4813]: I1206 15:48:41.264220 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 15:48:41 crc kubenswrapper[4813]: I1206 15:48:41.346515 4813 patch_prober.go:28] interesting pod/router-default-5444994796-ncvs7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 15:48:41 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 06 15:48:41 crc kubenswrapper[4813]: [+]process-running ok Dec 06 15:48:41 crc kubenswrapper[4813]: healthz check failed Dec 06 15:48:41 crc kubenswrapper[4813]: I1206 15:48:41.346573 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncvs7" podUID="8ddf8368-8232-4211-bfca-642b1acef6a4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 15:48:41 crc kubenswrapper[4813]: I1206 15:48:41.436484 4813 patch_prober.go:28] interesting pod/downloads-7954f5f757-kkbqh container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Dec 06 15:48:41 crc kubenswrapper[4813]: I1206 15:48:41.436531 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-kkbqh" podUID="26bc2fa2-9eee-4f7b-b31b-86f364df2b06" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Dec 06 15:48:41 crc kubenswrapper[4813]: I1206 15:48:41.437160 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3ac08880-2cdb-4dc7-a62a-d68e23e7919b-kubelet-dir\") pod \"3ac08880-2cdb-4dc7-a62a-d68e23e7919b\" (UID: \"3ac08880-2cdb-4dc7-a62a-d68e23e7919b\") " Dec 06 15:48:41 crc kubenswrapper[4813]: I1206 15:48:41.437292 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3ac08880-2cdb-4dc7-a62a-d68e23e7919b-kube-api-access\") pod \"3ac08880-2cdb-4dc7-a62a-d68e23e7919b\" (UID: \"3ac08880-2cdb-4dc7-a62a-d68e23e7919b\") " Dec 06 15:48:41 crc kubenswrapper[4813]: I1206 15:48:41.437661 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3ac08880-2cdb-4dc7-a62a-d68e23e7919b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "3ac08880-2cdb-4dc7-a62a-d68e23e7919b" (UID: "3ac08880-2cdb-4dc7-a62a-d68e23e7919b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 15:48:41 crc kubenswrapper[4813]: I1206 15:48:41.438724 4813 patch_prober.go:28] interesting pod/downloads-7954f5f757-kkbqh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Dec 06 15:48:41 crc kubenswrapper[4813]: I1206 15:48:41.438750 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-kkbqh" podUID="26bc2fa2-9eee-4f7b-b31b-86f364df2b06" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Dec 06 15:48:41 crc kubenswrapper[4813]: I1206 15:48:41.443092 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ac08880-2cdb-4dc7-a62a-d68e23e7919b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "3ac08880-2cdb-4dc7-a62a-d68e23e7919b" (UID: "3ac08880-2cdb-4dc7-a62a-d68e23e7919b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:48:41 crc kubenswrapper[4813]: I1206 15:48:41.546893 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3ac08880-2cdb-4dc7-a62a-d68e23e7919b-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 15:48:41 crc kubenswrapper[4813]: I1206 15:48:41.546932 4813 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3ac08880-2cdb-4dc7-a62a-d68e23e7919b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 15:48:41 crc kubenswrapper[4813]: I1206 15:48:41.598391 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-289lx" Dec 06 15:48:41 crc kubenswrapper[4813]: I1206 15:48:41.784820 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"3ac08880-2cdb-4dc7-a62a-d68e23e7919b","Type":"ContainerDied","Data":"54cfc28490e555375c706660e61a1dc12dca009bbc9b5a12b18cf8ff5d7a998e"} Dec 06 15:48:41 crc kubenswrapper[4813]: I1206 15:48:41.784861 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54cfc28490e555375c706660e61a1dc12dca009bbc9b5a12b18cf8ff5d7a998e" Dec 06 15:48:41 crc kubenswrapper[4813]: I1206 15:48:41.784880 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 15:48:42 crc kubenswrapper[4813]: I1206 15:48:42.347371 4813 patch_prober.go:28] interesting pod/router-default-5444994796-ncvs7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 15:48:42 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 06 15:48:42 crc kubenswrapper[4813]: [+]process-running ok Dec 06 15:48:42 crc kubenswrapper[4813]: healthz check failed Dec 06 15:48:42 crc kubenswrapper[4813]: I1206 15:48:42.347637 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncvs7" podUID="8ddf8368-8232-4211-bfca-642b1acef6a4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 15:48:43 crc kubenswrapper[4813]: I1206 15:48:43.351162 4813 patch_prober.go:28] interesting pod/router-default-5444994796-ncvs7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 15:48:43 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 06 15:48:43 crc kubenswrapper[4813]: [+]process-running ok Dec 06 15:48:43 crc kubenswrapper[4813]: healthz check failed Dec 06 15:48:43 crc kubenswrapper[4813]: I1206 15:48:43.351214 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncvs7" podUID="8ddf8368-8232-4211-bfca-642b1acef6a4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 15:48:43 crc kubenswrapper[4813]: I1206 15:48:43.473888 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18a94103-5ed7-4e27-a58d-7c989b0f70f3-metrics-certs\") pod \"network-metrics-daemon-kzbhb\" (UID: \"18a94103-5ed7-4e27-a58d-7c989b0f70f3\") " pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:48:43 crc kubenswrapper[4813]: I1206 15:48:43.478129 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18a94103-5ed7-4e27-a58d-7c989b0f70f3-metrics-certs\") pod \"network-metrics-daemon-kzbhb\" (UID: \"18a94103-5ed7-4e27-a58d-7c989b0f70f3\") " pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:48:43 crc kubenswrapper[4813]: I1206 15:48:43.511241 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kzbhb" Dec 06 15:48:44 crc kubenswrapper[4813]: I1206 15:48:44.348038 4813 patch_prober.go:28] interesting pod/router-default-5444994796-ncvs7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 15:48:44 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 06 15:48:44 crc kubenswrapper[4813]: [+]process-running ok Dec 06 15:48:44 crc kubenswrapper[4813]: healthz check failed Dec 06 15:48:44 crc kubenswrapper[4813]: I1206 15:48:44.348346 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncvs7" podUID="8ddf8368-8232-4211-bfca-642b1acef6a4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 15:48:44 crc kubenswrapper[4813]: I1206 15:48:44.546306 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-kzbhb"] Dec 06 15:48:44 crc kubenswrapper[4813]: W1206 15:48:44.584462 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18a94103_5ed7_4e27_a58d_7c989b0f70f3.slice/crio-4526f47f9b31ea5d31c64104b4989f0008c9f946cbb82819c41411f472b7f69d WatchSource:0}: Error finding container 4526f47f9b31ea5d31c64104b4989f0008c9f946cbb82819c41411f472b7f69d: Status 404 returned error can't find the container with id 4526f47f9b31ea5d31c64104b4989f0008c9f946cbb82819c41411f472b7f69d Dec 06 15:48:44 crc kubenswrapper[4813]: I1206 15:48:44.915620 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-kzbhb" event={"ID":"18a94103-5ed7-4e27-a58d-7c989b0f70f3","Type":"ContainerStarted","Data":"4526f47f9b31ea5d31c64104b4989f0008c9f946cbb82819c41411f472b7f69d"} Dec 06 15:48:45 crc kubenswrapper[4813]: I1206 15:48:45.346026 4813 patch_prober.go:28] interesting pod/router-default-5444994796-ncvs7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 15:48:45 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 06 15:48:45 crc kubenswrapper[4813]: [+]process-running ok Dec 06 15:48:45 crc kubenswrapper[4813]: healthz check failed Dec 06 15:48:45 crc kubenswrapper[4813]: I1206 15:48:45.346099 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncvs7" podUID="8ddf8368-8232-4211-bfca-642b1acef6a4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 15:48:46 crc kubenswrapper[4813]: I1206 15:48:46.349071 4813 patch_prober.go:28] interesting pod/router-default-5444994796-ncvs7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 15:48:46 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 06 15:48:46 crc kubenswrapper[4813]: [+]process-running ok Dec 06 15:48:46 crc kubenswrapper[4813]: healthz check failed Dec 06 15:48:46 crc kubenswrapper[4813]: I1206 15:48:46.349144 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncvs7" podUID="8ddf8368-8232-4211-bfca-642b1acef6a4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 15:48:47 crc kubenswrapper[4813]: I1206 15:48:47.347367 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-ncvs7" Dec 06 15:48:47 crc kubenswrapper[4813]: I1206 15:48:47.350128 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-ncvs7" Dec 06 15:48:47 crc kubenswrapper[4813]: I1206 15:48:47.939924 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-kzbhb" event={"ID":"18a94103-5ed7-4e27-a58d-7c989b0f70f3","Type":"ContainerStarted","Data":"084d2845e106fa199cf58ce4bf8ac9f39d9b23325fb23384e06e29275ae6c326"} Dec 06 15:48:48 crc kubenswrapper[4813]: I1206 15:48:48.965164 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-kzbhb" event={"ID":"18a94103-5ed7-4e27-a58d-7c989b0f70f3","Type":"ContainerStarted","Data":"56c8465903c931c958111ffae32567b2ac3eccae4c854e283b23e6567fb10d04"} Dec 06 15:48:48 crc kubenswrapper[4813]: I1206 15:48:48.983042 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-kzbhb" podStartSLOduration=148.983021438 podStartE2EDuration="2m28.983021438s" podCreationTimestamp="2025-12-06 15:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:48:48.982277658 +0000 UTC m=+168.873157234" watchObservedRunningTime="2025-12-06 15:48:48.983021438 +0000 UTC m=+168.873901014" Dec 06 15:48:49 crc kubenswrapper[4813]: I1206 15:48:49.437134 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 15:48:49 crc kubenswrapper[4813]: I1206 15:48:49.437203 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 15:48:50 crc kubenswrapper[4813]: I1206 15:48:50.514771 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-gb6cc" Dec 06 15:48:50 crc kubenswrapper[4813]: I1206 15:48:50.518468 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-gb6cc" Dec 06 15:48:51 crc kubenswrapper[4813]: I1206 15:48:51.436057 4813 patch_prober.go:28] interesting pod/downloads-7954f5f757-kkbqh container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Dec 06 15:48:51 crc kubenswrapper[4813]: I1206 15:48:51.436138 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-kkbqh" podUID="26bc2fa2-9eee-4f7b-b31b-86f364df2b06" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Dec 06 15:48:51 crc kubenswrapper[4813]: I1206 15:48:51.436194 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-kkbqh" Dec 06 15:48:51 crc kubenswrapper[4813]: I1206 15:48:51.436229 4813 patch_prober.go:28] interesting pod/downloads-7954f5f757-kkbqh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Dec 06 15:48:51 crc kubenswrapper[4813]: I1206 15:48:51.436318 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-kkbqh" podUID="26bc2fa2-9eee-4f7b-b31b-86f364df2b06" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Dec 06 15:48:51 crc kubenswrapper[4813]: I1206 15:48:51.436945 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"f89c7bb5a3731d5694812efe5ca0028f56d9afab1eb2aef5e5d8762e852f402b"} pod="openshift-console/downloads-7954f5f757-kkbqh" containerMessage="Container download-server failed liveness probe, will be restarted" Dec 06 15:48:51 crc kubenswrapper[4813]: I1206 15:48:51.437048 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-kkbqh" podUID="26bc2fa2-9eee-4f7b-b31b-86f364df2b06" containerName="download-server" containerID="cri-o://f89c7bb5a3731d5694812efe5ca0028f56d9afab1eb2aef5e5d8762e852f402b" gracePeriod=2 Dec 06 15:48:51 crc kubenswrapper[4813]: I1206 15:48:51.437336 4813 patch_prober.go:28] interesting pod/downloads-7954f5f757-kkbqh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Dec 06 15:48:51 crc kubenswrapper[4813]: I1206 15:48:51.437389 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-kkbqh" podUID="26bc2fa2-9eee-4f7b-b31b-86f364df2b06" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Dec 06 15:48:53 crc kubenswrapper[4813]: I1206 15:48:53.207919 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:49:01 crc kubenswrapper[4813]: I1206 15:49:01.437042 4813 patch_prober.go:28] interesting pod/downloads-7954f5f757-kkbqh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Dec 06 15:49:01 crc kubenswrapper[4813]: I1206 15:49:01.437894 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-kkbqh" podUID="26bc2fa2-9eee-4f7b-b31b-86f364df2b06" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Dec 06 15:49:01 crc kubenswrapper[4813]: I1206 15:49:01.549648 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fq7x6" Dec 06 15:49:05 crc kubenswrapper[4813]: I1206 15:49:05.332685 4813 generic.go:334] "Generic (PLEG): container finished" podID="26bc2fa2-9eee-4f7b-b31b-86f364df2b06" containerID="f89c7bb5a3731d5694812efe5ca0028f56d9afab1eb2aef5e5d8762e852f402b" exitCode=0 Dec 06 15:49:05 crc kubenswrapper[4813]: I1206 15:49:05.332784 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-kkbqh" event={"ID":"26bc2fa2-9eee-4f7b-b31b-86f364df2b06","Type":"ContainerDied","Data":"f89c7bb5a3731d5694812efe5ca0028f56d9afab1eb2aef5e5d8762e852f402b"} Dec 06 15:49:07 crc kubenswrapper[4813]: I1206 15:49:07.751124 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 15:49:11 crc kubenswrapper[4813]: I1206 15:49:11.436029 4813 patch_prober.go:28] interesting pod/downloads-7954f5f757-kkbqh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Dec 06 15:49:11 crc kubenswrapper[4813]: I1206 15:49:11.436328 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-kkbqh" podUID="26bc2fa2-9eee-4f7b-b31b-86f364df2b06" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Dec 06 15:49:11 crc kubenswrapper[4813]: I1206 15:49:11.613711 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 06 15:49:11 crc kubenswrapper[4813]: E1206 15:49:11.614006 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ac08880-2cdb-4dc7-a62a-d68e23e7919b" containerName="pruner" Dec 06 15:49:11 crc kubenswrapper[4813]: I1206 15:49:11.614017 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ac08880-2cdb-4dc7-a62a-d68e23e7919b" containerName="pruner" Dec 06 15:49:11 crc kubenswrapper[4813]: E1206 15:49:11.614030 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="977c2e53-21f6-4e10-a539-d57bcb28a6bb" containerName="pruner" Dec 06 15:49:11 crc kubenswrapper[4813]: I1206 15:49:11.614036 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="977c2e53-21f6-4e10-a539-d57bcb28a6bb" containerName="pruner" Dec 06 15:49:11 crc kubenswrapper[4813]: E1206 15:49:11.614046 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8810f12c-6dbc-4bf9-b27e-29ebc5986955" containerName="collect-profiles" Dec 06 15:49:11 crc kubenswrapper[4813]: I1206 15:49:11.614052 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="8810f12c-6dbc-4bf9-b27e-29ebc5986955" containerName="collect-profiles" Dec 06 15:49:11 crc kubenswrapper[4813]: I1206 15:49:11.614145 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ac08880-2cdb-4dc7-a62a-d68e23e7919b" containerName="pruner" Dec 06 15:49:11 crc kubenswrapper[4813]: I1206 15:49:11.614154 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="977c2e53-21f6-4e10-a539-d57bcb28a6bb" containerName="pruner" Dec 06 15:49:11 crc kubenswrapper[4813]: I1206 15:49:11.614162 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="8810f12c-6dbc-4bf9-b27e-29ebc5986955" containerName="collect-profiles" Dec 06 15:49:11 crc kubenswrapper[4813]: I1206 15:49:11.614591 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 15:49:11 crc kubenswrapper[4813]: I1206 15:49:11.617246 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 06 15:49:11 crc kubenswrapper[4813]: I1206 15:49:11.620761 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 06 15:49:11 crc kubenswrapper[4813]: I1206 15:49:11.620945 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 06 15:49:11 crc kubenswrapper[4813]: I1206 15:49:11.777799 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/70ee83ca-e4e1-46a9-a16d-0897aa3e1349-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"70ee83ca-e4e1-46a9-a16d-0897aa3e1349\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 15:49:11 crc kubenswrapper[4813]: I1206 15:49:11.777904 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/70ee83ca-e4e1-46a9-a16d-0897aa3e1349-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"70ee83ca-e4e1-46a9-a16d-0897aa3e1349\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 15:49:11 crc kubenswrapper[4813]: I1206 15:49:11.879063 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/70ee83ca-e4e1-46a9-a16d-0897aa3e1349-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"70ee83ca-e4e1-46a9-a16d-0897aa3e1349\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 15:49:11 crc kubenswrapper[4813]: I1206 15:49:11.879405 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/70ee83ca-e4e1-46a9-a16d-0897aa3e1349-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"70ee83ca-e4e1-46a9-a16d-0897aa3e1349\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 15:49:11 crc kubenswrapper[4813]: I1206 15:49:11.880560 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/70ee83ca-e4e1-46a9-a16d-0897aa3e1349-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"70ee83ca-e4e1-46a9-a16d-0897aa3e1349\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 15:49:11 crc kubenswrapper[4813]: I1206 15:49:11.910456 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/70ee83ca-e4e1-46a9-a16d-0897aa3e1349-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"70ee83ca-e4e1-46a9-a16d-0897aa3e1349\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 15:49:11 crc kubenswrapper[4813]: I1206 15:49:11.939961 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 15:49:15 crc kubenswrapper[4813]: I1206 15:49:15.976499 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fszk7"] Dec 06 15:49:16 crc kubenswrapper[4813]: I1206 15:49:16.600352 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 06 15:49:16 crc kubenswrapper[4813]: I1206 15:49:16.601252 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 15:49:16 crc kubenswrapper[4813]: I1206 15:49:16.612227 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 06 15:49:16 crc kubenswrapper[4813]: I1206 15:49:16.638405 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff-kubelet-dir\") pod \"installer-9-crc\" (UID: \"6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 15:49:16 crc kubenswrapper[4813]: I1206 15:49:16.638437 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff-var-lock\") pod \"installer-9-crc\" (UID: \"6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 15:49:16 crc kubenswrapper[4813]: I1206 15:49:16.638456 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff-kube-api-access\") pod \"installer-9-crc\" (UID: \"6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 15:49:16 crc kubenswrapper[4813]: I1206 15:49:16.739932 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff-kubelet-dir\") pod \"installer-9-crc\" (UID: \"6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 15:49:16 crc kubenswrapper[4813]: I1206 15:49:16.739981 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff-kube-api-access\") pod \"installer-9-crc\" (UID: \"6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 15:49:16 crc kubenswrapper[4813]: I1206 15:49:16.740000 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff-var-lock\") pod \"installer-9-crc\" (UID: \"6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 15:49:16 crc kubenswrapper[4813]: I1206 15:49:16.740038 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff-kubelet-dir\") pod \"installer-9-crc\" (UID: \"6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 15:49:16 crc kubenswrapper[4813]: I1206 15:49:16.740090 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff-var-lock\") pod \"installer-9-crc\" (UID: \"6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 15:49:16 crc kubenswrapper[4813]: I1206 15:49:16.756418 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff-kube-api-access\") pod \"installer-9-crc\" (UID: \"6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 15:49:16 crc kubenswrapper[4813]: I1206 15:49:16.930200 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 15:49:19 crc kubenswrapper[4813]: I1206 15:49:19.427922 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 15:49:19 crc kubenswrapper[4813]: I1206 15:49:19.428901 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 15:49:19 crc kubenswrapper[4813]: E1206 15:49:19.901434 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 06 15:49:19 crc kubenswrapper[4813]: E1206 15:49:19.901649 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pqpc5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-7d6jw_openshift-marketplace(952e1f68-97e3-4191-9dde-43d67e83a1e0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 15:49:19 crc kubenswrapper[4813]: E1206 15:49:19.902840 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-7d6jw" podUID="952e1f68-97e3-4191-9dde-43d67e83a1e0" Dec 06 15:49:21 crc kubenswrapper[4813]: E1206 15:49:21.382211 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-7d6jw" podUID="952e1f68-97e3-4191-9dde-43d67e83a1e0" Dec 06 15:49:21 crc kubenswrapper[4813]: I1206 15:49:21.437592 4813 patch_prober.go:28] interesting pod/downloads-7954f5f757-kkbqh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Dec 06 15:49:21 crc kubenswrapper[4813]: I1206 15:49:21.437833 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-kkbqh" podUID="26bc2fa2-9eee-4f7b-b31b-86f364df2b06" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Dec 06 15:49:21 crc kubenswrapper[4813]: E1206 15:49:21.441797 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 06 15:49:21 crc kubenswrapper[4813]: E1206 15:49:21.441933 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nqpfc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-6rwpr_openshift-marketplace(76073738-387a-4968-a483-8880e866d1e4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 15:49:21 crc kubenswrapper[4813]: E1206 15:49:21.443144 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-6rwpr" podUID="76073738-387a-4968-a483-8880e866d1e4" Dec 06 15:49:24 crc kubenswrapper[4813]: E1206 15:49:24.407653 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-6rwpr" podUID="76073738-387a-4968-a483-8880e866d1e4" Dec 06 15:49:24 crc kubenswrapper[4813]: E1206 15:49:24.561490 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 06 15:49:24 crc kubenswrapper[4813]: E1206 15:49:24.561640 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vpc2w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-kp6t2_openshift-marketplace(e982bb16-5f82-418b-8c03-8c73d22e010b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 15:49:24 crc kubenswrapper[4813]: E1206 15:49:24.562800 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-kp6t2" podUID="e982bb16-5f82-418b-8c03-8c73d22e010b" Dec 06 15:49:24 crc kubenswrapper[4813]: E1206 15:49:24.872344 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 06 15:49:24 crc kubenswrapper[4813]: E1206 15:49:24.872672 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kv5hk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-5zp6r_openshift-marketplace(6a553ef5-b76a-46cd-9e49-30bf4ca86e29): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 15:49:24 crc kubenswrapper[4813]: E1206 15:49:24.873857 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-5zp6r" podUID="6a553ef5-b76a-46cd-9e49-30bf4ca86e29" Dec 06 15:49:24 crc kubenswrapper[4813]: I1206 15:49:24.919763 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 06 15:49:24 crc kubenswrapper[4813]: W1206 15:49:24.929464 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod70ee83ca_e4e1_46a9_a16d_0897aa3e1349.slice/crio-f1b11df7d193350ea97498bb3f48b97d2d6fd6f8608416e81a4f44e09a3702a7 WatchSource:0}: Error finding container f1b11df7d193350ea97498bb3f48b97d2d6fd6f8608416e81a4f44e09a3702a7: Status 404 returned error can't find the container with id f1b11df7d193350ea97498bb3f48b97d2d6fd6f8608416e81a4f44e09a3702a7 Dec 06 15:49:25 crc kubenswrapper[4813]: I1206 15:49:25.084491 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 06 15:49:25 crc kubenswrapper[4813]: W1206 15:49:25.103622 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod6ea9fcd6_1a07_4903_a615_3a2e2d01c6ff.slice/crio-18ef8c6d8028f565b48f2a37f4a2210eac51883a2d7100dbc27b823509599a37 WatchSource:0}: Error finding container 18ef8c6d8028f565b48f2a37f4a2210eac51883a2d7100dbc27b823509599a37: Status 404 returned error can't find the container with id 18ef8c6d8028f565b48f2a37f4a2210eac51883a2d7100dbc27b823509599a37 Dec 06 15:49:25 crc kubenswrapper[4813]: I1206 15:49:25.452820 4813 generic.go:334] "Generic (PLEG): container finished" podID="7e2017ee-a77f-47b3-ae48-058c62ce30d6" containerID="21757bd7d17a48fab7ad6737cf23ac70341558e2105fdd30d90055dccccdf160" exitCode=0 Dec 06 15:49:25 crc kubenswrapper[4813]: I1206 15:49:25.453012 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pvphv" event={"ID":"7e2017ee-a77f-47b3-ae48-058c62ce30d6","Type":"ContainerDied","Data":"21757bd7d17a48fab7ad6737cf23ac70341558e2105fdd30d90055dccccdf160"} Dec 06 15:49:25 crc kubenswrapper[4813]: I1206 15:49:25.455702 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"70ee83ca-e4e1-46a9-a16d-0897aa3e1349","Type":"ContainerStarted","Data":"f1b11df7d193350ea97498bb3f48b97d2d6fd6f8608416e81a4f44e09a3702a7"} Dec 06 15:49:25 crc kubenswrapper[4813]: I1206 15:49:25.457404 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vn7fl" event={"ID":"da4de6a4-9efc-425c-8712-e66919babdf2","Type":"ContainerStarted","Data":"27bbb547810dbc893ccfc765f55472dc563d743bb19c743f7e039356935dc817"} Dec 06 15:49:25 crc kubenswrapper[4813]: I1206 15:49:25.459716 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-kkbqh" event={"ID":"26bc2fa2-9eee-4f7b-b31b-86f364df2b06","Type":"ContainerStarted","Data":"e9ec8f55e274543cd4c769136099a5499f0c98910de6e4079697d772fe46e286"} Dec 06 15:49:25 crc kubenswrapper[4813]: I1206 15:49:25.461763 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff","Type":"ContainerStarted","Data":"18ef8c6d8028f565b48f2a37f4a2210eac51883a2d7100dbc27b823509599a37"} Dec 06 15:49:25 crc kubenswrapper[4813]: E1206 15:49:25.462742 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-kp6t2" podUID="e982bb16-5f82-418b-8c03-8c73d22e010b" Dec 06 15:49:25 crc kubenswrapper[4813]: E1206 15:49:25.462985 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-5zp6r" podUID="6a553ef5-b76a-46cd-9e49-30bf4ca86e29" Dec 06 15:49:26 crc kubenswrapper[4813]: I1206 15:49:26.466325 4813 generic.go:334] "Generic (PLEG): container finished" podID="70ee83ca-e4e1-46a9-a16d-0897aa3e1349" containerID="88525a3a57ee5ebe60e501c4493d730805e954bdc054b5a403518c8afc7d5376" exitCode=0 Dec 06 15:49:26 crc kubenswrapper[4813]: I1206 15:49:26.467274 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"70ee83ca-e4e1-46a9-a16d-0897aa3e1349","Type":"ContainerDied","Data":"88525a3a57ee5ebe60e501c4493d730805e954bdc054b5a403518c8afc7d5376"} Dec 06 15:49:26 crc kubenswrapper[4813]: I1206 15:49:26.468859 4813 generic.go:334] "Generic (PLEG): container finished" podID="da4de6a4-9efc-425c-8712-e66919babdf2" containerID="27bbb547810dbc893ccfc765f55472dc563d743bb19c743f7e039356935dc817" exitCode=0 Dec 06 15:49:26 crc kubenswrapper[4813]: I1206 15:49:26.468959 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vn7fl" event={"ID":"da4de6a4-9efc-425c-8712-e66919babdf2","Type":"ContainerDied","Data":"27bbb547810dbc893ccfc765f55472dc563d743bb19c743f7e039356935dc817"} Dec 06 15:49:26 crc kubenswrapper[4813]: I1206 15:49:26.473675 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff","Type":"ContainerStarted","Data":"ae4d733b020f8e38c80a6329efd041cf5fb29b01887560ad571f5fb1be515359"} Dec 06 15:49:26 crc kubenswrapper[4813]: I1206 15:49:26.473718 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-kkbqh" Dec 06 15:49:26 crc kubenswrapper[4813]: I1206 15:49:26.474228 4813 patch_prober.go:28] interesting pod/downloads-7954f5f757-kkbqh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Dec 06 15:49:26 crc kubenswrapper[4813]: I1206 15:49:26.474458 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-kkbqh" podUID="26bc2fa2-9eee-4f7b-b31b-86f364df2b06" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Dec 06 15:49:26 crc kubenswrapper[4813]: I1206 15:49:26.507803 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=10.507781307 podStartE2EDuration="10.507781307s" podCreationTimestamp="2025-12-06 15:49:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:49:26.506120201 +0000 UTC m=+206.396999797" watchObservedRunningTime="2025-12-06 15:49:26.507781307 +0000 UTC m=+206.398660913" Dec 06 15:49:26 crc kubenswrapper[4813]: E1206 15:49:26.535991 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 06 15:49:26 crc kubenswrapper[4813]: E1206 15:49:26.536275 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-59zbh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-qnq74_openshift-marketplace(f522807f-660d-4791-95f4-f2d9daa58936): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 15:49:26 crc kubenswrapper[4813]: E1206 15:49:26.537508 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-qnq74" podUID="f522807f-660d-4791-95f4-f2d9daa58936" Dec 06 15:49:27 crc kubenswrapper[4813]: I1206 15:49:27.479429 4813 patch_prober.go:28] interesting pod/downloads-7954f5f757-kkbqh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Dec 06 15:49:27 crc kubenswrapper[4813]: I1206 15:49:27.479480 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-kkbqh" podUID="26bc2fa2-9eee-4f7b-b31b-86f364df2b06" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Dec 06 15:49:27 crc kubenswrapper[4813]: E1206 15:49:27.484465 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-qnq74" podUID="f522807f-660d-4791-95f4-f2d9daa58936" Dec 06 15:49:27 crc kubenswrapper[4813]: I1206 15:49:27.769223 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 15:49:27 crc kubenswrapper[4813]: I1206 15:49:27.912748 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/70ee83ca-e4e1-46a9-a16d-0897aa3e1349-kubelet-dir\") pod \"70ee83ca-e4e1-46a9-a16d-0897aa3e1349\" (UID: \"70ee83ca-e4e1-46a9-a16d-0897aa3e1349\") " Dec 06 15:49:27 crc kubenswrapper[4813]: I1206 15:49:27.912825 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/70ee83ca-e4e1-46a9-a16d-0897aa3e1349-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "70ee83ca-e4e1-46a9-a16d-0897aa3e1349" (UID: "70ee83ca-e4e1-46a9-a16d-0897aa3e1349"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 15:49:27 crc kubenswrapper[4813]: I1206 15:49:27.912852 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/70ee83ca-e4e1-46a9-a16d-0897aa3e1349-kube-api-access\") pod \"70ee83ca-e4e1-46a9-a16d-0897aa3e1349\" (UID: \"70ee83ca-e4e1-46a9-a16d-0897aa3e1349\") " Dec 06 15:49:27 crc kubenswrapper[4813]: I1206 15:49:27.913183 4813 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/70ee83ca-e4e1-46a9-a16d-0897aa3e1349-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 15:49:27 crc kubenswrapper[4813]: I1206 15:49:27.928524 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70ee83ca-e4e1-46a9-a16d-0897aa3e1349-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "70ee83ca-e4e1-46a9-a16d-0897aa3e1349" (UID: "70ee83ca-e4e1-46a9-a16d-0897aa3e1349"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:49:28 crc kubenswrapper[4813]: I1206 15:49:28.014772 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/70ee83ca-e4e1-46a9-a16d-0897aa3e1349-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 15:49:28 crc kubenswrapper[4813]: E1206 15:49:28.326904 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 06 15:49:28 crc kubenswrapper[4813]: E1206 15:49:28.327059 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k59xj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-54xzm_openshift-marketplace(e2fac5b3-c60b-4c96-8841-ae53a24aaed1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 15:49:28 crc kubenswrapper[4813]: E1206 15:49:28.328908 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-54xzm" podUID="e2fac5b3-c60b-4c96-8841-ae53a24aaed1" Dec 06 15:49:28 crc kubenswrapper[4813]: I1206 15:49:28.485036 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 15:49:28 crc kubenswrapper[4813]: I1206 15:49:28.488312 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"70ee83ca-e4e1-46a9-a16d-0897aa3e1349","Type":"ContainerDied","Data":"f1b11df7d193350ea97498bb3f48b97d2d6fd6f8608416e81a4f44e09a3702a7"} Dec 06 15:49:28 crc kubenswrapper[4813]: I1206 15:49:28.488341 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1b11df7d193350ea97498bb3f48b97d2d6fd6f8608416e81a4f44e09a3702a7" Dec 06 15:49:28 crc kubenswrapper[4813]: E1206 15:49:28.488421 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-54xzm" podUID="e2fac5b3-c60b-4c96-8841-ae53a24aaed1" Dec 06 15:49:31 crc kubenswrapper[4813]: I1206 15:49:31.436322 4813 patch_prober.go:28] interesting pod/downloads-7954f5f757-kkbqh container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Dec 06 15:49:31 crc kubenswrapper[4813]: I1206 15:49:31.437989 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-kkbqh" podUID="26bc2fa2-9eee-4f7b-b31b-86f364df2b06" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Dec 06 15:49:31 crc kubenswrapper[4813]: I1206 15:49:31.439884 4813 patch_prober.go:28] interesting pod/downloads-7954f5f757-kkbqh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Dec 06 15:49:31 crc kubenswrapper[4813]: I1206 15:49:31.439990 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-kkbqh" podUID="26bc2fa2-9eee-4f7b-b31b-86f364df2b06" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Dec 06 15:49:31 crc kubenswrapper[4813]: I1206 15:49:31.501090 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vn7fl" event={"ID":"da4de6a4-9efc-425c-8712-e66919babdf2","Type":"ContainerStarted","Data":"6041b776a12e9bbefe8831d00231faeec17ad0e52b7045f06b8132aa5d850c9c"} Dec 06 15:49:32 crc kubenswrapper[4813]: I1206 15:49:32.543813 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vn7fl" podStartSLOduration=4.423438221 podStartE2EDuration="59.543797411s" podCreationTimestamp="2025-12-06 15:48:33 +0000 UTC" firstStartedPulling="2025-12-06 15:48:35.550308403 +0000 UTC m=+155.441187979" lastFinishedPulling="2025-12-06 15:49:30.670667583 +0000 UTC m=+210.561547169" observedRunningTime="2025-12-06 15:49:32.543466442 +0000 UTC m=+212.434346018" watchObservedRunningTime="2025-12-06 15:49:32.543797411 +0000 UTC m=+212.434676987" Dec 06 15:49:34 crc kubenswrapper[4813]: I1206 15:49:34.035672 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vn7fl" Dec 06 15:49:34 crc kubenswrapper[4813]: I1206 15:49:34.036041 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vn7fl" Dec 06 15:49:34 crc kubenswrapper[4813]: I1206 15:49:34.533686 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pvphv" event={"ID":"7e2017ee-a77f-47b3-ae48-058c62ce30d6","Type":"ContainerStarted","Data":"f585663920f0680a9df968c5c2db672e4401bd52d2731b387c809ae6b8415530"} Dec 06 15:49:34 crc kubenswrapper[4813]: I1206 15:49:34.562513 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pvphv" podStartSLOduration=5.294351576 podStartE2EDuration="1m1.562495868s" podCreationTimestamp="2025-12-06 15:48:33 +0000 UTC" firstStartedPulling="2025-12-06 15:48:36.585939505 +0000 UTC m=+156.476819081" lastFinishedPulling="2025-12-06 15:49:32.854083797 +0000 UTC m=+212.744963373" observedRunningTime="2025-12-06 15:49:34.554497417 +0000 UTC m=+214.445377023" watchObservedRunningTime="2025-12-06 15:49:34.562495868 +0000 UTC m=+214.453375444" Dec 06 15:49:35 crc kubenswrapper[4813]: I1206 15:49:35.437186 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-vn7fl" podUID="da4de6a4-9efc-425c-8712-e66919babdf2" containerName="registry-server" probeResult="failure" output=< Dec 06 15:49:35 crc kubenswrapper[4813]: timeout: failed to connect service ":50051" within 1s Dec 06 15:49:35 crc kubenswrapper[4813]: > Dec 06 15:49:40 crc kubenswrapper[4813]: I1206 15:49:40.599803 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rwpr" event={"ID":"76073738-387a-4968-a483-8880e866d1e4","Type":"ContainerStarted","Data":"cd856a72ba6a857e1551e16c291ffcd37c2f23cf257cb4aceaa7b977c234eef8"} Dec 06 15:49:40 crc kubenswrapper[4813]: I1206 15:49:40.602338 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7d6jw" event={"ID":"952e1f68-97e3-4191-9dde-43d67e83a1e0","Type":"ContainerStarted","Data":"5c912bf5b91f301bd9b0351d051dfb46642aa95008de5ccb5206464e915fc5e5"} Dec 06 15:49:41 crc kubenswrapper[4813]: I1206 15:49:41.026128 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" podUID="994c44ca-f520-4452-b267-961bd400908e" containerName="oauth-openshift" containerID="cri-o://032b09a0f6c8db36ef202f022f145601133f8c4a49b4551eec4e58d2758c9caa" gracePeriod=15 Dec 06 15:49:41 crc kubenswrapper[4813]: I1206 15:49:41.461892 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-kkbqh" Dec 06 15:49:41 crc kubenswrapper[4813]: I1206 15:49:41.609474 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5zp6r" event={"ID":"6a553ef5-b76a-46cd-9e49-30bf4ca86e29","Type":"ContainerStarted","Data":"de8103356b82043f938ff194c779e0f049cb25beacc418f3491629dc29cadf88"} Dec 06 15:49:41 crc kubenswrapper[4813]: I1206 15:49:41.612408 4813 generic.go:334] "Generic (PLEG): container finished" podID="994c44ca-f520-4452-b267-961bd400908e" containerID="032b09a0f6c8db36ef202f022f145601133f8c4a49b4551eec4e58d2758c9caa" exitCode=0 Dec 06 15:49:41 crc kubenswrapper[4813]: I1206 15:49:41.612499 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" event={"ID":"994c44ca-f520-4452-b267-961bd400908e","Type":"ContainerDied","Data":"032b09a0f6c8db36ef202f022f145601133f8c4a49b4551eec4e58d2758c9caa"} Dec 06 15:49:41 crc kubenswrapper[4813]: I1206 15:49:41.614071 4813 generic.go:334] "Generic (PLEG): container finished" podID="76073738-387a-4968-a483-8880e866d1e4" containerID="cd856a72ba6a857e1551e16c291ffcd37c2f23cf257cb4aceaa7b977c234eef8" exitCode=0 Dec 06 15:49:41 crc kubenswrapper[4813]: I1206 15:49:41.614112 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rwpr" event={"ID":"76073738-387a-4968-a483-8880e866d1e4","Type":"ContainerDied","Data":"cd856a72ba6a857e1551e16c291ffcd37c2f23cf257cb4aceaa7b977c234eef8"} Dec 06 15:49:44 crc kubenswrapper[4813]: I1206 15:49:44.227629 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vn7fl" Dec 06 15:49:44 crc kubenswrapper[4813]: I1206 15:49:44.302571 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vn7fl" Dec 06 15:49:44 crc kubenswrapper[4813]: I1206 15:49:44.307716 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pvphv" Dec 06 15:49:44 crc kubenswrapper[4813]: I1206 15:49:44.307792 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pvphv" Dec 06 15:49:44 crc kubenswrapper[4813]: I1206 15:49:44.379378 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pvphv" Dec 06 15:49:44 crc kubenswrapper[4813]: I1206 15:49:44.633602 4813 generic.go:334] "Generic (PLEG): container finished" podID="6a553ef5-b76a-46cd-9e49-30bf4ca86e29" containerID="de8103356b82043f938ff194c779e0f049cb25beacc418f3491629dc29cadf88" exitCode=0 Dec 06 15:49:44 crc kubenswrapper[4813]: I1206 15:49:44.633691 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5zp6r" event={"ID":"6a553ef5-b76a-46cd-9e49-30bf4ca86e29","Type":"ContainerDied","Data":"de8103356b82043f938ff194c779e0f049cb25beacc418f3491629dc29cadf88"} Dec 06 15:49:44 crc kubenswrapper[4813]: I1206 15:49:44.638907 4813 generic.go:334] "Generic (PLEG): container finished" podID="952e1f68-97e3-4191-9dde-43d67e83a1e0" containerID="5c912bf5b91f301bd9b0351d051dfb46642aa95008de5ccb5206464e915fc5e5" exitCode=0 Dec 06 15:49:44 crc kubenswrapper[4813]: I1206 15:49:44.639077 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7d6jw" event={"ID":"952e1f68-97e3-4191-9dde-43d67e83a1e0","Type":"ContainerDied","Data":"5c912bf5b91f301bd9b0351d051dfb46642aa95008de5ccb5206464e915fc5e5"} Dec 06 15:49:44 crc kubenswrapper[4813]: I1206 15:49:44.726171 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pvphv" Dec 06 15:49:46 crc kubenswrapper[4813]: I1206 15:49:46.954898 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pvphv"] Dec 06 15:49:46 crc kubenswrapper[4813]: I1206 15:49:46.955380 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pvphv" podUID="7e2017ee-a77f-47b3-ae48-058c62ce30d6" containerName="registry-server" containerID="cri-o://f585663920f0680a9df968c5c2db672e4401bd52d2731b387c809ae6b8415530" gracePeriod=2 Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.602094 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.615314 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/994c44ca-f520-4452-b267-961bd400908e-audit-dir\") pod \"994c44ca-f520-4452-b267-961bd400908e\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.615394 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-user-idp-0-file-data\") pod \"994c44ca-f520-4452-b267-961bd400908e\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.615437 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-cliconfig\") pod \"994c44ca-f520-4452-b267-961bd400908e\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.615494 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-user-template-error\") pod \"994c44ca-f520-4452-b267-961bd400908e\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.615531 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-router-certs\") pod \"994c44ca-f520-4452-b267-961bd400908e\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.615569 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8b5mh\" (UniqueName: \"kubernetes.io/projected/994c44ca-f520-4452-b267-961bd400908e-kube-api-access-8b5mh\") pod \"994c44ca-f520-4452-b267-961bd400908e\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.615605 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/994c44ca-f520-4452-b267-961bd400908e-audit-policies\") pod \"994c44ca-f520-4452-b267-961bd400908e\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.615646 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-user-template-login\") pod \"994c44ca-f520-4452-b267-961bd400908e\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.615683 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-ocp-branding-template\") pod \"994c44ca-f520-4452-b267-961bd400908e\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.615746 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-serving-cert\") pod \"994c44ca-f520-4452-b267-961bd400908e\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.615787 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-session\") pod \"994c44ca-f520-4452-b267-961bd400908e\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.615830 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-service-ca\") pod \"994c44ca-f520-4452-b267-961bd400908e\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.615868 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-user-template-provider-selection\") pod \"994c44ca-f520-4452-b267-961bd400908e\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.615907 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-trusted-ca-bundle\") pod \"994c44ca-f520-4452-b267-961bd400908e\" (UID: \"994c44ca-f520-4452-b267-961bd400908e\") " Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.616790 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/994c44ca-f520-4452-b267-961bd400908e-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "994c44ca-f520-4452-b267-961bd400908e" (UID: "994c44ca-f520-4452-b267-961bd400908e"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.617737 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "994c44ca-f520-4452-b267-961bd400908e" (UID: "994c44ca-f520-4452-b267-961bd400908e"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.618769 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/994c44ca-f520-4452-b267-961bd400908e-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "994c44ca-f520-4452-b267-961bd400908e" (UID: "994c44ca-f520-4452-b267-961bd400908e"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.620830 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "994c44ca-f520-4452-b267-961bd400908e" (UID: "994c44ca-f520-4452-b267-961bd400908e"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.621220 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "994c44ca-f520-4452-b267-961bd400908e" (UID: "994c44ca-f520-4452-b267-961bd400908e"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.626303 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "994c44ca-f520-4452-b267-961bd400908e" (UID: "994c44ca-f520-4452-b267-961bd400908e"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.628178 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "994c44ca-f520-4452-b267-961bd400908e" (UID: "994c44ca-f520-4452-b267-961bd400908e"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.636884 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "994c44ca-f520-4452-b267-961bd400908e" (UID: "994c44ca-f520-4452-b267-961bd400908e"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.636387 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "994c44ca-f520-4452-b267-961bd400908e" (UID: "994c44ca-f520-4452-b267-961bd400908e"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.638701 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "994c44ca-f520-4452-b267-961bd400908e" (UID: "994c44ca-f520-4452-b267-961bd400908e"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.644852 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "994c44ca-f520-4452-b267-961bd400908e" (UID: "994c44ca-f520-4452-b267-961bd400908e"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.645220 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "994c44ca-f520-4452-b267-961bd400908e" (UID: "994c44ca-f520-4452-b267-961bd400908e"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.649512 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/994c44ca-f520-4452-b267-961bd400908e-kube-api-access-8b5mh" (OuterVolumeSpecName: "kube-api-access-8b5mh") pod "994c44ca-f520-4452-b267-961bd400908e" (UID: "994c44ca-f520-4452-b267-961bd400908e"). InnerVolumeSpecName "kube-api-access-8b5mh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.667344 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "994c44ca-f520-4452-b267-961bd400908e" (UID: "994c44ca-f520-4452-b267-961bd400908e"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.672382 4813 generic.go:334] "Generic (PLEG): container finished" podID="7e2017ee-a77f-47b3-ae48-058c62ce30d6" containerID="f585663920f0680a9df968c5c2db672e4401bd52d2731b387c809ae6b8415530" exitCode=0 Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.673067 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pvphv" event={"ID":"7e2017ee-a77f-47b3-ae48-058c62ce30d6","Type":"ContainerDied","Data":"f585663920f0680a9df968c5c2db672e4401bd52d2731b387c809ae6b8415530"} Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.674691 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" event={"ID":"994c44ca-f520-4452-b267-961bd400908e","Type":"ContainerDied","Data":"016f963155111634ee3a14bfc168a30594f6053f636547169d7e28b46ff766a7"} Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.674732 4813 scope.go:117] "RemoveContainer" containerID="032b09a0f6c8db36ef202f022f145601133f8c4a49b4551eec4e58d2758c9caa" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.674901 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fszk7" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.690651 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-d646c55fb-s7trl"] Dec 06 15:49:47 crc kubenswrapper[4813]: E1206 15:49:47.691351 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="994c44ca-f520-4452-b267-961bd400908e" containerName="oauth-openshift" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.691422 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="994c44ca-f520-4452-b267-961bd400908e" containerName="oauth-openshift" Dec 06 15:49:47 crc kubenswrapper[4813]: E1206 15:49:47.691458 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70ee83ca-e4e1-46a9-a16d-0897aa3e1349" containerName="pruner" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.691466 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="70ee83ca-e4e1-46a9-a16d-0897aa3e1349" containerName="pruner" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.691606 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="994c44ca-f520-4452-b267-961bd400908e" containerName="oauth-openshift" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.691630 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="70ee83ca-e4e1-46a9-a16d-0897aa3e1349" containerName="pruner" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.692541 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.693759 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-d646c55fb-s7trl"] Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.695751 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.697592 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.697861 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.698158 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.698370 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.698532 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.698673 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.699191 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.699393 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.699763 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.699776 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.699952 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.702440 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.711821 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.717162 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.717401 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.717416 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8b5mh\" (UniqueName: \"kubernetes.io/projected/994c44ca-f520-4452-b267-961bd400908e-kube-api-access-8b5mh\") on node \"crc\" DevicePath \"\"" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.717426 4813 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/994c44ca-f520-4452-b267-961bd400908e-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.717437 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.717446 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.717455 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.717774 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.717792 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.717802 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.717814 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.717822 4813 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/994c44ca-f520-4452-b267-961bd400908e-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.717831 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.717841 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/994c44ca-f520-4452-b267-961bd400908e-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.722443 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.742230 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fszk7"] Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.745813 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fszk7"] Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.819124 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.819453 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.819493 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-system-router-certs\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.819520 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.819538 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-system-service-ca\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.819555 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/315ab197-8861-4e58-b296-1f0ff794c6d1-audit-policies\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.819573 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-system-session\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.819587 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/315ab197-8861-4e58-b296-1f0ff794c6d1-audit-dir\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.819622 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.819638 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.819655 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-user-template-error\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.819675 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-user-template-login\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.819691 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.819708 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmh7n\" (UniqueName: \"kubernetes.io/projected/315ab197-8861-4e58-b296-1f0ff794c6d1-kube-api-access-mmh7n\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.920457 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-user-template-login\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.920498 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.920524 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmh7n\" (UniqueName: \"kubernetes.io/projected/315ab197-8861-4e58-b296-1f0ff794c6d1-kube-api-access-mmh7n\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.920569 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.920589 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.920611 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-system-router-certs\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.920632 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.920647 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/315ab197-8861-4e58-b296-1f0ff794c6d1-audit-policies\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.920664 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-system-service-ca\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.920680 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/315ab197-8861-4e58-b296-1f0ff794c6d1-audit-dir\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.920695 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-system-session\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.920731 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.920746 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.920764 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-user-template-error\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.921471 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/315ab197-8861-4e58-b296-1f0ff794c6d1-audit-dir\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.922080 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/315ab197-8861-4e58-b296-1f0ff794c6d1-audit-policies\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.922525 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.922545 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-system-service-ca\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.923453 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.926613 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-user-template-error\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.926727 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-user-template-login\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.927043 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.927479 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-system-router-certs\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.927847 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.929484 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.929801 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-system-session\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.930572 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/315ab197-8861-4e58-b296-1f0ff794c6d1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:47 crc kubenswrapper[4813]: I1206 15:49:47.947157 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmh7n\" (UniqueName: \"kubernetes.io/projected/315ab197-8861-4e58-b296-1f0ff794c6d1-kube-api-access-mmh7n\") pod \"oauth-openshift-d646c55fb-s7trl\" (UID: \"315ab197-8861-4e58-b296-1f0ff794c6d1\") " pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:48 crc kubenswrapper[4813]: I1206 15:49:48.018118 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:48 crc kubenswrapper[4813]: I1206 15:49:48.377696 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pvphv" Dec 06 15:49:48 crc kubenswrapper[4813]: I1206 15:49:48.494211 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="994c44ca-f520-4452-b267-961bd400908e" path="/var/lib/kubelet/pods/994c44ca-f520-4452-b267-961bd400908e/volumes" Dec 06 15:49:48 crc kubenswrapper[4813]: I1206 15:49:48.528504 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e2017ee-a77f-47b3-ae48-058c62ce30d6-catalog-content\") pod \"7e2017ee-a77f-47b3-ae48-058c62ce30d6\" (UID: \"7e2017ee-a77f-47b3-ae48-058c62ce30d6\") " Dec 06 15:49:48 crc kubenswrapper[4813]: I1206 15:49:48.528750 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e2017ee-a77f-47b3-ae48-058c62ce30d6-utilities\") pod \"7e2017ee-a77f-47b3-ae48-058c62ce30d6\" (UID: \"7e2017ee-a77f-47b3-ae48-058c62ce30d6\") " Dec 06 15:49:48 crc kubenswrapper[4813]: I1206 15:49:48.528852 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-445mw\" (UniqueName: \"kubernetes.io/projected/7e2017ee-a77f-47b3-ae48-058c62ce30d6-kube-api-access-445mw\") pod \"7e2017ee-a77f-47b3-ae48-058c62ce30d6\" (UID: \"7e2017ee-a77f-47b3-ae48-058c62ce30d6\") " Dec 06 15:49:48 crc kubenswrapper[4813]: I1206 15:49:48.529600 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e2017ee-a77f-47b3-ae48-058c62ce30d6-utilities" (OuterVolumeSpecName: "utilities") pod "7e2017ee-a77f-47b3-ae48-058c62ce30d6" (UID: "7e2017ee-a77f-47b3-ae48-058c62ce30d6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:49:48 crc kubenswrapper[4813]: I1206 15:49:48.531825 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e2017ee-a77f-47b3-ae48-058c62ce30d6-kube-api-access-445mw" (OuterVolumeSpecName: "kube-api-access-445mw") pod "7e2017ee-a77f-47b3-ae48-058c62ce30d6" (UID: "7e2017ee-a77f-47b3-ae48-058c62ce30d6"). InnerVolumeSpecName "kube-api-access-445mw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:49:48 crc kubenswrapper[4813]: I1206 15:49:48.548601 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e2017ee-a77f-47b3-ae48-058c62ce30d6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7e2017ee-a77f-47b3-ae48-058c62ce30d6" (UID: "7e2017ee-a77f-47b3-ae48-058c62ce30d6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:49:48 crc kubenswrapper[4813]: I1206 15:49:48.630580 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e2017ee-a77f-47b3-ae48-058c62ce30d6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 15:49:48 crc kubenswrapper[4813]: I1206 15:49:48.630838 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e2017ee-a77f-47b3-ae48-058c62ce30d6-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 15:49:48 crc kubenswrapper[4813]: I1206 15:49:48.630954 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-445mw\" (UniqueName: \"kubernetes.io/projected/7e2017ee-a77f-47b3-ae48-058c62ce30d6-kube-api-access-445mw\") on node \"crc\" DevicePath \"\"" Dec 06 15:49:48 crc kubenswrapper[4813]: I1206 15:49:48.682184 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pvphv" event={"ID":"7e2017ee-a77f-47b3-ae48-058c62ce30d6","Type":"ContainerDied","Data":"9dc3b473db2f7f0e842d40c093a40b6cc5f7da306114acbfe2632a40859abbe0"} Dec 06 15:49:48 crc kubenswrapper[4813]: I1206 15:49:48.682576 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pvphv" Dec 06 15:49:48 crc kubenswrapper[4813]: I1206 15:49:48.711499 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pvphv"] Dec 06 15:49:48 crc kubenswrapper[4813]: I1206 15:49:48.714755 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pvphv"] Dec 06 15:49:49 crc kubenswrapper[4813]: I1206 15:49:49.427367 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 15:49:49 crc kubenswrapper[4813]: I1206 15:49:49.427456 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 15:49:49 crc kubenswrapper[4813]: I1206 15:49:49.427500 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" Dec 06 15:49:49 crc kubenswrapper[4813]: I1206 15:49:49.428086 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624"} pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 15:49:49 crc kubenswrapper[4813]: I1206 15:49:49.428134 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" containerID="cri-o://0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624" gracePeriod=600 Dec 06 15:49:49 crc kubenswrapper[4813]: I1206 15:49:49.755104 4813 scope.go:117] "RemoveContainer" containerID="f585663920f0680a9df968c5c2db672e4401bd52d2731b387c809ae6b8415530" Dec 06 15:49:50 crc kubenswrapper[4813]: I1206 15:49:50.498400 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e2017ee-a77f-47b3-ae48-058c62ce30d6" path="/var/lib/kubelet/pods/7e2017ee-a77f-47b3-ae48-058c62ce30d6/volumes" Dec 06 15:49:50 crc kubenswrapper[4813]: I1206 15:49:50.632614 4813 scope.go:117] "RemoveContainer" containerID="21757bd7d17a48fab7ad6737cf23ac70341558e2105fdd30d90055dccccdf160" Dec 06 15:49:50 crc kubenswrapper[4813]: I1206 15:49:50.681228 4813 scope.go:117] "RemoveContainer" containerID="ff2d29a7f28dd730b01271bf394ad2bbb513ece6aa593c67a8793e32f5e407bd" Dec 06 15:49:50 crc kubenswrapper[4813]: I1206 15:49:50.704875 4813 generic.go:334] "Generic (PLEG): container finished" podID="d88e8bae-c055-4c55-b548-f621ff96de06" containerID="0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624" exitCode=0 Dec 06 15:49:50 crc kubenswrapper[4813]: I1206 15:49:50.705027 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerDied","Data":"0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624"} Dec 06 15:49:51 crc kubenswrapper[4813]: I1206 15:49:51.328838 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-d646c55fb-s7trl"] Dec 06 15:49:51 crc kubenswrapper[4813]: W1206 15:49:51.341217 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod315ab197_8861_4e58_b296_1f0ff794c6d1.slice/crio-01f82b5d27edc34c9b46c74d1f3ed69f3a5bee3f4af02514dda3ca298e5a4b62 WatchSource:0}: Error finding container 01f82b5d27edc34c9b46c74d1f3ed69f3a5bee3f4af02514dda3ca298e5a4b62: Status 404 returned error can't find the container with id 01f82b5d27edc34c9b46c74d1f3ed69f3a5bee3f4af02514dda3ca298e5a4b62 Dec 06 15:49:51 crc kubenswrapper[4813]: I1206 15:49:51.719928 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qnq74" event={"ID":"f522807f-660d-4791-95f4-f2d9daa58936","Type":"ContainerStarted","Data":"98c3136dd68e10777f5a954d2dc9f8a9e069ef68281dc21b5388a3a6125bf467"} Dec 06 15:49:51 crc kubenswrapper[4813]: I1206 15:49:51.722808 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5zp6r" event={"ID":"6a553ef5-b76a-46cd-9e49-30bf4ca86e29","Type":"ContainerStarted","Data":"5ff03ddb8621464af047b3a4ddbde8a63a7fd6cf544c0e8e25ca07066341a555"} Dec 06 15:49:51 crc kubenswrapper[4813]: I1206 15:49:51.724472 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerStarted","Data":"9bb3ab0a5b69e0227b05bdb28844955eb8df2d6874e1937ca3d3672f21c852e9"} Dec 06 15:49:51 crc kubenswrapper[4813]: I1206 15:49:51.726382 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" event={"ID":"315ab197-8861-4e58-b296-1f0ff794c6d1","Type":"ContainerStarted","Data":"ad193fad45b460ab0833a09b90b1d0a88ce6433d4f47392f83ace8697f57e8af"} Dec 06 15:49:51 crc kubenswrapper[4813]: I1206 15:49:51.726412 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" event={"ID":"315ab197-8861-4e58-b296-1f0ff794c6d1","Type":"ContainerStarted","Data":"01f82b5d27edc34c9b46c74d1f3ed69f3a5bee3f4af02514dda3ca298e5a4b62"} Dec 06 15:49:51 crc kubenswrapper[4813]: I1206 15:49:51.727048 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:51 crc kubenswrapper[4813]: I1206 15:49:51.730701 4813 patch_prober.go:28] interesting pod/oauth-openshift-d646c55fb-s7trl container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.56:6443/healthz\": dial tcp 10.217.0.56:6443: connect: connection refused" start-of-body= Dec 06 15:49:51 crc kubenswrapper[4813]: I1206 15:49:51.730740 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" podUID="315ab197-8861-4e58-b296-1f0ff794c6d1" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.56:6443/healthz\": dial tcp 10.217.0.56:6443: connect: connection refused" Dec 06 15:49:51 crc kubenswrapper[4813]: I1206 15:49:51.739629 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7d6jw" event={"ID":"952e1f68-97e3-4191-9dde-43d67e83a1e0","Type":"ContainerStarted","Data":"fc5c2d08cc45d223b3dddb5282d706c77ec8854aa99c6a85001fec025351536f"} Dec 06 15:49:51 crc kubenswrapper[4813]: I1206 15:49:51.745329 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rwpr" event={"ID":"76073738-387a-4968-a483-8880e866d1e4","Type":"ContainerStarted","Data":"dc210c0684fff6a7e1dbf1d2841de30e46e582fb956d958e6cb6ed48c0837da8"} Dec 06 15:49:51 crc kubenswrapper[4813]: I1206 15:49:51.748782 4813 generic.go:334] "Generic (PLEG): container finished" podID="e2fac5b3-c60b-4c96-8841-ae53a24aaed1" containerID="1bb1bb72773c10356ac5e2a9f3183c87652f6b8c98b8b7524252529c7813f497" exitCode=0 Dec 06 15:49:51 crc kubenswrapper[4813]: I1206 15:49:51.748860 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-54xzm" event={"ID":"e2fac5b3-c60b-4c96-8841-ae53a24aaed1","Type":"ContainerDied","Data":"1bb1bb72773c10356ac5e2a9f3183c87652f6b8c98b8b7524252529c7813f497"} Dec 06 15:49:51 crc kubenswrapper[4813]: I1206 15:49:51.755303 4813 generic.go:334] "Generic (PLEG): container finished" podID="e982bb16-5f82-418b-8c03-8c73d22e010b" containerID="009b5e53cb626741235644fa6abe27c1d2b476b93cdd1de42ae8b6e277c6d5bb" exitCode=0 Dec 06 15:49:51 crc kubenswrapper[4813]: I1206 15:49:51.755348 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kp6t2" event={"ID":"e982bb16-5f82-418b-8c03-8c73d22e010b","Type":"ContainerDied","Data":"009b5e53cb626741235644fa6abe27c1d2b476b93cdd1de42ae8b6e277c6d5bb"} Dec 06 15:49:51 crc kubenswrapper[4813]: I1206 15:49:51.762158 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5zp6r" podStartSLOduration=3.464215062 podStartE2EDuration="1m18.759968787s" podCreationTimestamp="2025-12-06 15:48:33 +0000 UTC" firstStartedPulling="2025-12-06 15:48:35.574783501 +0000 UTC m=+155.465663077" lastFinishedPulling="2025-12-06 15:49:50.870537216 +0000 UTC m=+230.761416802" observedRunningTime="2025-12-06 15:49:51.758377573 +0000 UTC m=+231.649257149" watchObservedRunningTime="2025-12-06 15:49:51.759968787 +0000 UTC m=+231.650848363" Dec 06 15:49:51 crc kubenswrapper[4813]: I1206 15:49:51.780418 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" podStartSLOduration=35.780401802 podStartE2EDuration="35.780401802s" podCreationTimestamp="2025-12-06 15:49:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:49:51.7803329 +0000 UTC m=+231.671212476" watchObservedRunningTime="2025-12-06 15:49:51.780401802 +0000 UTC m=+231.671281378" Dec 06 15:49:51 crc kubenswrapper[4813]: I1206 15:49:51.838019 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7d6jw" podStartSLOduration=2.502837701 podStartE2EDuration="1m17.838001996s" podCreationTimestamp="2025-12-06 15:48:34 +0000 UTC" firstStartedPulling="2025-12-06 15:48:35.540536904 +0000 UTC m=+155.431416480" lastFinishedPulling="2025-12-06 15:49:50.875701199 +0000 UTC m=+230.766580775" observedRunningTime="2025-12-06 15:49:51.837768959 +0000 UTC m=+231.728648535" watchObservedRunningTime="2025-12-06 15:49:51.838001996 +0000 UTC m=+231.728881572" Dec 06 15:49:51 crc kubenswrapper[4813]: I1206 15:49:51.854247 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6rwpr" podStartSLOduration=3.4198771949999998 podStartE2EDuration="1m20.854231835s" podCreationTimestamp="2025-12-06 15:48:31 +0000 UTC" firstStartedPulling="2025-12-06 15:48:33.260977328 +0000 UTC m=+153.151856904" lastFinishedPulling="2025-12-06 15:49:50.695331928 +0000 UTC m=+230.586211544" observedRunningTime="2025-12-06 15:49:51.852776344 +0000 UTC m=+231.743655920" watchObservedRunningTime="2025-12-06 15:49:51.854231835 +0000 UTC m=+231.745111411" Dec 06 15:49:52 crc kubenswrapper[4813]: I1206 15:49:52.762865 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-54xzm" event={"ID":"e2fac5b3-c60b-4c96-8841-ae53a24aaed1","Type":"ContainerStarted","Data":"58ab538dbec20a657dfd0e11d156896a425cf1cf03ac50cbbb65c4454607e6f2"} Dec 06 15:49:52 crc kubenswrapper[4813]: I1206 15:49:52.766796 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kp6t2" event={"ID":"e982bb16-5f82-418b-8c03-8c73d22e010b","Type":"ContainerStarted","Data":"6e4cbadc5a42dd6f512afa67014f67d6c1c96d3dbe1b330a18894f6c5d98b8d3"} Dec 06 15:49:52 crc kubenswrapper[4813]: I1206 15:49:52.768436 4813 generic.go:334] "Generic (PLEG): container finished" podID="f522807f-660d-4791-95f4-f2d9daa58936" containerID="98c3136dd68e10777f5a954d2dc9f8a9e069ef68281dc21b5388a3a6125bf467" exitCode=0 Dec 06 15:49:52 crc kubenswrapper[4813]: I1206 15:49:52.768532 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qnq74" event={"ID":"f522807f-660d-4791-95f4-f2d9daa58936","Type":"ContainerDied","Data":"98c3136dd68e10777f5a954d2dc9f8a9e069ef68281dc21b5388a3a6125bf467"} Dec 06 15:49:52 crc kubenswrapper[4813]: I1206 15:49:52.772788 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-d646c55fb-s7trl" Dec 06 15:49:52 crc kubenswrapper[4813]: I1206 15:49:52.784374 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-54xzm" podStartSLOduration=3.871816503 podStartE2EDuration="1m22.784360171s" podCreationTimestamp="2025-12-06 15:48:30 +0000 UTC" firstStartedPulling="2025-12-06 15:48:33.249276768 +0000 UTC m=+153.140156344" lastFinishedPulling="2025-12-06 15:49:52.161820436 +0000 UTC m=+232.052700012" observedRunningTime="2025-12-06 15:49:52.782426408 +0000 UTC m=+232.673305984" watchObservedRunningTime="2025-12-06 15:49:52.784360171 +0000 UTC m=+232.675239747" Dec 06 15:49:52 crc kubenswrapper[4813]: I1206 15:49:52.829878 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kp6t2" podStartSLOduration=4.244608902 podStartE2EDuration="1m21.82986015s" podCreationTimestamp="2025-12-06 15:48:31 +0000 UTC" firstStartedPulling="2025-12-06 15:48:34.535990445 +0000 UTC m=+154.426870021" lastFinishedPulling="2025-12-06 15:49:52.121241693 +0000 UTC m=+232.012121269" observedRunningTime="2025-12-06 15:49:52.803504661 +0000 UTC m=+232.694384247" watchObservedRunningTime="2025-12-06 15:49:52.82986015 +0000 UTC m=+232.720739726" Dec 06 15:49:53 crc kubenswrapper[4813]: I1206 15:49:53.774984 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qnq74" event={"ID":"f522807f-660d-4791-95f4-f2d9daa58936","Type":"ContainerStarted","Data":"ab86ee25c416c62178029b32a6c3639723f9b985501d0ac91049782cf3d230be"} Dec 06 15:49:53 crc kubenswrapper[4813]: I1206 15:49:53.794710 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qnq74" podStartSLOduration=2.906442253 podStartE2EDuration="1m22.794692026s" podCreationTimestamp="2025-12-06 15:48:31 +0000 UTC" firstStartedPulling="2025-12-06 15:48:33.256304564 +0000 UTC m=+153.147184140" lastFinishedPulling="2025-12-06 15:49:53.144554337 +0000 UTC m=+233.035433913" observedRunningTime="2025-12-06 15:49:53.793972306 +0000 UTC m=+233.684851882" watchObservedRunningTime="2025-12-06 15:49:53.794692026 +0000 UTC m=+233.685571592" Dec 06 15:49:54 crc kubenswrapper[4813]: I1206 15:49:54.112949 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5zp6r" Dec 06 15:49:54 crc kubenswrapper[4813]: I1206 15:49:54.112995 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5zp6r" Dec 06 15:49:54 crc kubenswrapper[4813]: I1206 15:49:54.520066 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7d6jw" Dec 06 15:49:54 crc kubenswrapper[4813]: I1206 15:49:54.520459 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7d6jw" Dec 06 15:49:55 crc kubenswrapper[4813]: I1206 15:49:55.560106 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7d6jw" podUID="952e1f68-97e3-4191-9dde-43d67e83a1e0" containerName="registry-server" probeResult="failure" output=< Dec 06 15:49:55 crc kubenswrapper[4813]: timeout: failed to connect service ":50051" within 1s Dec 06 15:49:55 crc kubenswrapper[4813]: > Dec 06 15:49:55 crc kubenswrapper[4813]: I1206 15:49:55.567583 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5zp6r" podUID="6a553ef5-b76a-46cd-9e49-30bf4ca86e29" containerName="registry-server" probeResult="failure" output=< Dec 06 15:49:55 crc kubenswrapper[4813]: timeout: failed to connect service ":50051" within 1s Dec 06 15:49:55 crc kubenswrapper[4813]: > Dec 06 15:50:00 crc kubenswrapper[4813]: I1206 15:50:00.908332 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-54xzm" Dec 06 15:50:00 crc kubenswrapper[4813]: I1206 15:50:00.909976 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-54xzm" Dec 06 15:50:00 crc kubenswrapper[4813]: I1206 15:50:00.947967 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-54xzm" Dec 06 15:50:01 crc kubenswrapper[4813]: I1206 15:50:01.618059 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6rwpr" Dec 06 15:50:01 crc kubenswrapper[4813]: I1206 15:50:01.618402 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6rwpr" Dec 06 15:50:01 crc kubenswrapper[4813]: I1206 15:50:01.663252 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6rwpr" Dec 06 15:50:01 crc kubenswrapper[4813]: I1206 15:50:01.723091 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kp6t2" Dec 06 15:50:01 crc kubenswrapper[4813]: I1206 15:50:01.724540 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kp6t2" Dec 06 15:50:01 crc kubenswrapper[4813]: I1206 15:50:01.781483 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kp6t2" Dec 06 15:50:01 crc kubenswrapper[4813]: I1206 15:50:01.853186 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6rwpr" Dec 06 15:50:01 crc kubenswrapper[4813]: I1206 15:50:01.860628 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-54xzm" Dec 06 15:50:01 crc kubenswrapper[4813]: I1206 15:50:01.860786 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kp6t2" Dec 06 15:50:01 crc kubenswrapper[4813]: I1206 15:50:01.980787 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qnq74" Dec 06 15:50:01 crc kubenswrapper[4813]: I1206 15:50:01.980826 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qnq74" Dec 06 15:50:02 crc kubenswrapper[4813]: I1206 15:50:02.016672 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qnq74" Dec 06 15:50:02 crc kubenswrapper[4813]: I1206 15:50:02.854251 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qnq74" Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.804206 4813 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 06 15:50:03 crc kubenswrapper[4813]: E1206 15:50:03.804593 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e2017ee-a77f-47b3-ae48-058c62ce30d6" containerName="registry-server" Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.804615 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e2017ee-a77f-47b3-ae48-058c62ce30d6" containerName="registry-server" Dec 06 15:50:03 crc kubenswrapper[4813]: E1206 15:50:03.804643 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e2017ee-a77f-47b3-ae48-058c62ce30d6" containerName="extract-utilities" Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.804656 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e2017ee-a77f-47b3-ae48-058c62ce30d6" containerName="extract-utilities" Dec 06 15:50:03 crc kubenswrapper[4813]: E1206 15:50:03.804691 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e2017ee-a77f-47b3-ae48-058c62ce30d6" containerName="extract-content" Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.804703 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e2017ee-a77f-47b3-ae48-058c62ce30d6" containerName="extract-content" Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.804877 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e2017ee-a77f-47b3-ae48-058c62ce30d6" containerName="registry-server" Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.805466 4813 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.805615 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.805840 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137" gracePeriod=15 Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.805863 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1" gracePeriod=15 Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.806007 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789" gracePeriod=15 Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.806095 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0" gracePeriod=15 Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.806168 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb" gracePeriod=15 Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.807983 4813 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 15:50:03 crc kubenswrapper[4813]: E1206 15:50:03.808585 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.808602 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 06 15:50:03 crc kubenswrapper[4813]: E1206 15:50:03.808612 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.808619 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 15:50:03 crc kubenswrapper[4813]: E1206 15:50:03.808638 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.808644 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 06 15:50:03 crc kubenswrapper[4813]: E1206 15:50:03.808651 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.808657 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 06 15:50:03 crc kubenswrapper[4813]: E1206 15:50:03.808667 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.808673 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 06 15:50:03 crc kubenswrapper[4813]: E1206 15:50:03.808683 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.808689 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.808834 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.808850 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.808857 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.808863 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.808871 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.808878 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 15:50:03 crc kubenswrapper[4813]: E1206 15:50:03.809036 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.809048 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.935964 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.936030 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.936056 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.936069 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.936105 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.936124 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.936139 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 15:50:03 crc kubenswrapper[4813]: I1206 15:50:03.936188 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.002243 4813 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" start-of-body= Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.002308 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.036969 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.037219 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.037424 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.037529 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.037623 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.037746 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.037841 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.037969 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.038101 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.038227 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.038324 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.038415 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.038516 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.038613 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.038703 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.038793 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.165832 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5zp6r" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.167530 4813 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.167989 4813 status_manager.go:851] "Failed to get status for pod" podUID="6a553ef5-b76a-46cd-9e49-30bf4ca86e29" pod="openshift-marketplace/redhat-operators-5zp6r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-5zp6r\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.213689 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5zp6r" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.214425 4813 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.214807 4813 status_manager.go:851] "Failed to get status for pod" podUID="6a553ef5-b76a-46cd-9e49-30bf4ca86e29" pod="openshift-marketplace/redhat-operators-5zp6r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-5zp6r\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.561074 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7d6jw" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.561734 4813 status_manager.go:851] "Failed to get status for pod" podUID="6a553ef5-b76a-46cd-9e49-30bf4ca86e29" pod="openshift-marketplace/redhat-operators-5zp6r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-5zp6r\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.562335 4813 status_manager.go:851] "Failed to get status for pod" podUID="952e1f68-97e3-4191-9dde-43d67e83a1e0" pod="openshift-marketplace/redhat-operators-7d6jw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-7d6jw\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.600342 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7d6jw" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.600875 4813 status_manager.go:851] "Failed to get status for pod" podUID="6a553ef5-b76a-46cd-9e49-30bf4ca86e29" pod="openshift-marketplace/redhat-operators-5zp6r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-5zp6r\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.601337 4813 status_manager.go:851] "Failed to get status for pod" podUID="952e1f68-97e3-4191-9dde-43d67e83a1e0" pod="openshift-marketplace/redhat-operators-7d6jw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-7d6jw\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.836058 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.838026 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 15:50:04 crc kubenswrapper[4813]: I1206 15:50:04.839299 4813 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb" exitCode=2 Dec 06 15:50:05 crc kubenswrapper[4813]: I1206 15:50:05.846681 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 06 15:50:05 crc kubenswrapper[4813]: I1206 15:50:05.848274 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 15:50:05 crc kubenswrapper[4813]: I1206 15:50:05.849243 4813 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1" exitCode=0 Dec 06 15:50:05 crc kubenswrapper[4813]: I1206 15:50:05.849300 4813 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789" exitCode=0 Dec 06 15:50:05 crc kubenswrapper[4813]: I1206 15:50:05.849312 4813 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0" exitCode=0 Dec 06 15:50:05 crc kubenswrapper[4813]: I1206 15:50:05.849335 4813 scope.go:117] "RemoveContainer" containerID="20591f3d28021639bd2785c67847d11676580812c33c779d14f1d6f452479935" Dec 06 15:50:05 crc kubenswrapper[4813]: I1206 15:50:05.851360 4813 generic.go:334] "Generic (PLEG): container finished" podID="6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff" containerID="ae4d733b020f8e38c80a6329efd041cf5fb29b01887560ad571f5fb1be515359" exitCode=0 Dec 06 15:50:05 crc kubenswrapper[4813]: I1206 15:50:05.851444 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff","Type":"ContainerDied","Data":"ae4d733b020f8e38c80a6329efd041cf5fb29b01887560ad571f5fb1be515359"} Dec 06 15:50:05 crc kubenswrapper[4813]: I1206 15:50:05.852252 4813 status_manager.go:851] "Failed to get status for pod" podUID="6a553ef5-b76a-46cd-9e49-30bf4ca86e29" pod="openshift-marketplace/redhat-operators-5zp6r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-5zp6r\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:05 crc kubenswrapper[4813]: I1206 15:50:05.852612 4813 status_manager.go:851] "Failed to get status for pod" podUID="952e1f68-97e3-4191-9dde-43d67e83a1e0" pod="openshift-marketplace/redhat-operators-7d6jw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-7d6jw\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:05 crc kubenswrapper[4813]: I1206 15:50:05.852774 4813 status_manager.go:851] "Failed to get status for pod" podUID="6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.183817 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.185083 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.185537 4813 status_manager.go:851] "Failed to get status for pod" podUID="952e1f68-97e3-4191-9dde-43d67e83a1e0" pod="openshift-marketplace/redhat-operators-7d6jw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-7d6jw\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.185686 4813 status_manager.go:851] "Failed to get status for pod" podUID="6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.185821 4813 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.185958 4813 status_manager.go:851] "Failed to get status for pod" podUID="6a553ef5-b76a-46cd-9e49-30bf4ca86e29" pod="openshift-marketplace/redhat-operators-5zp6r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-5zp6r\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.281905 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.282003 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.282047 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.282055 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.282113 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.282196 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.282542 4813 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.282562 4813 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.282572 4813 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.492333 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.860836 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.862966 4813 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137" exitCode=0 Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.863252 4813 scope.go:117] "RemoveContainer" containerID="16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1" Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.863322 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.866322 4813 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.866693 4813 status_manager.go:851] "Failed to get status for pod" podUID="6a553ef5-b76a-46cd-9e49-30bf4ca86e29" pod="openshift-marketplace/redhat-operators-5zp6r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-5zp6r\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.867045 4813 status_manager.go:851] "Failed to get status for pod" podUID="952e1f68-97e3-4191-9dde-43d67e83a1e0" pod="openshift-marketplace/redhat-operators-7d6jw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-7d6jw\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.867574 4813 status_manager.go:851] "Failed to get status for pod" podUID="6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.868379 4813 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.868537 4813 status_manager.go:851] "Failed to get status for pod" podUID="6a553ef5-b76a-46cd-9e49-30bf4ca86e29" pod="openshift-marketplace/redhat-operators-5zp6r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-5zp6r\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.868808 4813 status_manager.go:851] "Failed to get status for pod" podUID="952e1f68-97e3-4191-9dde-43d67e83a1e0" pod="openshift-marketplace/redhat-operators-7d6jw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-7d6jw\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.869275 4813 status_manager.go:851] "Failed to get status for pod" podUID="6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.883671 4813 scope.go:117] "RemoveContainer" containerID="98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789" Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.903498 4813 scope.go:117] "RemoveContainer" containerID="74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0" Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.929354 4813 scope.go:117] "RemoveContainer" containerID="4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb" Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.964575 4813 scope.go:117] "RemoveContainer" containerID="3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137" Dec 06 15:50:06 crc kubenswrapper[4813]: I1206 15:50:06.989551 4813 scope.go:117] "RemoveContainer" containerID="5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7" Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.029876 4813 scope.go:117] "RemoveContainer" containerID="16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1" Dec 06 15:50:07 crc kubenswrapper[4813]: E1206 15:50:07.030800 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\": container with ID starting with 16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1 not found: ID does not exist" containerID="16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1" Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.030825 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1"} err="failed to get container status \"16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\": rpc error: code = NotFound desc = could not find container \"16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1\": container with ID starting with 16ebfae0c53e75eb49e7f0152d56f8b5c0696060c776349dd0de59b71e881df1 not found: ID does not exist" Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.030844 4813 scope.go:117] "RemoveContainer" containerID="98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789" Dec 06 15:50:07 crc kubenswrapper[4813]: E1206 15:50:07.031119 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\": container with ID starting with 98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789 not found: ID does not exist" containerID="98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789" Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.031155 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789"} err="failed to get container status \"98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\": rpc error: code = NotFound desc = could not find container \"98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789\": container with ID starting with 98c72507be8836b3a696b68ed91f6fda17ea5de4f6919444ce4a898290671789 not found: ID does not exist" Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.031182 4813 scope.go:117] "RemoveContainer" containerID="74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0" Dec 06 15:50:07 crc kubenswrapper[4813]: E1206 15:50:07.031626 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\": container with ID starting with 74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0 not found: ID does not exist" containerID="74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0" Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.031645 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0"} err="failed to get container status \"74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\": rpc error: code = NotFound desc = could not find container \"74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0\": container with ID starting with 74384f452ec5102e3bf4cf83074ad23a087420ef057c54e033eb151cb47658b0 not found: ID does not exist" Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.031936 4813 scope.go:117] "RemoveContainer" containerID="4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb" Dec 06 15:50:07 crc kubenswrapper[4813]: E1206 15:50:07.032254 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\": container with ID starting with 4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb not found: ID does not exist" containerID="4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb" Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.032361 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb"} err="failed to get container status \"4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\": rpc error: code = NotFound desc = could not find container \"4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb\": container with ID starting with 4725578abaa543b385a802ef0f47b19edf10651bce57c0c8b1bba2e988f487eb not found: ID does not exist" Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.032376 4813 scope.go:117] "RemoveContainer" containerID="3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137" Dec 06 15:50:07 crc kubenswrapper[4813]: E1206 15:50:07.032635 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\": container with ID starting with 3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137 not found: ID does not exist" containerID="3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137" Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.032673 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137"} err="failed to get container status \"3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\": rpc error: code = NotFound desc = could not find container \"3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137\": container with ID starting with 3244206196fe6e0c643900f4bb5eb3e4eff3f24fd4bdc313329770e45cd38137 not found: ID does not exist" Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.032689 4813 scope.go:117] "RemoveContainer" containerID="5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7" Dec 06 15:50:07 crc kubenswrapper[4813]: E1206 15:50:07.032936 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\": container with ID starting with 5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7 not found: ID does not exist" containerID="5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7" Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.032955 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7"} err="failed to get container status \"5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\": rpc error: code = NotFound desc = could not find container \"5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7\": container with ID starting with 5732e3ca3c97404f4a1dd453eb158d5111699399a548c074c2610a823c8f05d7 not found: ID does not exist" Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.133241 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.133674 4813 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.133969 4813 status_manager.go:851] "Failed to get status for pod" podUID="6a553ef5-b76a-46cd-9e49-30bf4ca86e29" pod="openshift-marketplace/redhat-operators-5zp6r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-5zp6r\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.134514 4813 status_manager.go:851] "Failed to get status for pod" podUID="952e1f68-97e3-4191-9dde-43d67e83a1e0" pod="openshift-marketplace/redhat-operators-7d6jw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-7d6jw\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.134735 4813 status_manager.go:851] "Failed to get status for pod" podUID="6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.293599 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff-var-lock\") pod \"6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff\" (UID: \"6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff\") " Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.293696 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff-var-lock" (OuterVolumeSpecName: "var-lock") pod "6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff" (UID: "6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.293726 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff-kubelet-dir\") pod \"6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff\" (UID: \"6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff\") " Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.293758 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff" (UID: "6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.293806 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff-kube-api-access\") pod \"6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff\" (UID: \"6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff\") " Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.294293 4813 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff-var-lock\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.294333 4813 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.299770 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff" (UID: "6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.395346 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.873815 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff","Type":"ContainerDied","Data":"18ef8c6d8028f565b48f2a37f4a2210eac51883a2d7100dbc27b823509599a37"} Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.873860 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="18ef8c6d8028f565b48f2a37f4a2210eac51883a2d7100dbc27b823509599a37" Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.873938 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.887013 4813 status_manager.go:851] "Failed to get status for pod" podUID="6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.887340 4813 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.887770 4813 status_manager.go:851] "Failed to get status for pod" podUID="6a553ef5-b76a-46cd-9e49-30bf4ca86e29" pod="openshift-marketplace/redhat-operators-5zp6r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-5zp6r\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:07 crc kubenswrapper[4813]: I1206 15:50:07.887975 4813 status_manager.go:851] "Failed to get status for pod" podUID="952e1f68-97e3-4191-9dde-43d67e83a1e0" pod="openshift-marketplace/redhat-operators-7d6jw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-7d6jw\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:08 crc kubenswrapper[4813]: E1206 15:50:08.837322 4813 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.9:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 15:50:08 crc kubenswrapper[4813]: I1206 15:50:08.837752 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 15:50:08 crc kubenswrapper[4813]: E1206 15:50:08.866889 4813 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.9:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187eab149529c9ad openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 15:50:08.866052525 +0000 UTC m=+248.756932111,LastTimestamp:2025-12-06 15:50:08.866052525 +0000 UTC m=+248.756932111,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 15:50:08 crc kubenswrapper[4813]: I1206 15:50:08.889246 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"7ac95e6a297efa229e1ad30e01bd8f9d8d351889f6fd209362fa64b0313c9a4a"} Dec 06 15:50:09 crc kubenswrapper[4813]: I1206 15:50:09.897098 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"0eb54412ce9438f87b8eb1721001f4d5bb20f02f364dc78f126722417786f8ac"} Dec 06 15:50:09 crc kubenswrapper[4813]: I1206 15:50:09.897833 4813 status_manager.go:851] "Failed to get status for pod" podUID="6a553ef5-b76a-46cd-9e49-30bf4ca86e29" pod="openshift-marketplace/redhat-operators-5zp6r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-5zp6r\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:09 crc kubenswrapper[4813]: I1206 15:50:09.898140 4813 status_manager.go:851] "Failed to get status for pod" podUID="952e1f68-97e3-4191-9dde-43d67e83a1e0" pod="openshift-marketplace/redhat-operators-7d6jw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-7d6jw\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:09 crc kubenswrapper[4813]: E1206 15:50:09.898180 4813 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.9:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 15:50:09 crc kubenswrapper[4813]: I1206 15:50:09.898496 4813 status_manager.go:851] "Failed to get status for pod" podUID="6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:10 crc kubenswrapper[4813]: E1206 15:50:10.393976 4813 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:10 crc kubenswrapper[4813]: E1206 15:50:10.394956 4813 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:10 crc kubenswrapper[4813]: E1206 15:50:10.395976 4813 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:10 crc kubenswrapper[4813]: E1206 15:50:10.396546 4813 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:10 crc kubenswrapper[4813]: E1206 15:50:10.397009 4813 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:10 crc kubenswrapper[4813]: I1206 15:50:10.397053 4813 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 06 15:50:10 crc kubenswrapper[4813]: E1206 15:50:10.397660 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="200ms" Dec 06 15:50:10 crc kubenswrapper[4813]: I1206 15:50:10.491324 4813 status_manager.go:851] "Failed to get status for pod" podUID="6a553ef5-b76a-46cd-9e49-30bf4ca86e29" pod="openshift-marketplace/redhat-operators-5zp6r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-5zp6r\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:10 crc kubenswrapper[4813]: I1206 15:50:10.491676 4813 status_manager.go:851] "Failed to get status for pod" podUID="952e1f68-97e3-4191-9dde-43d67e83a1e0" pod="openshift-marketplace/redhat-operators-7d6jw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-7d6jw\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:10 crc kubenswrapper[4813]: I1206 15:50:10.492048 4813 status_manager.go:851] "Failed to get status for pod" podUID="6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:10 crc kubenswrapper[4813]: E1206 15:50:10.599280 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="400ms" Dec 06 15:50:10 crc kubenswrapper[4813]: E1206 15:50:10.902663 4813 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.9:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 15:50:11 crc kubenswrapper[4813]: E1206 15:50:11.000643 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="800ms" Dec 06 15:50:11 crc kubenswrapper[4813]: E1206 15:50:11.801728 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="1.6s" Dec 06 15:50:13 crc kubenswrapper[4813]: E1206 15:50:13.403447 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="3.2s" Dec 06 15:50:16 crc kubenswrapper[4813]: E1206 15:50:16.343140 4813 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.9:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187eab149529c9ad openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 15:50:08.866052525 +0000 UTC m=+248.756932111,LastTimestamp:2025-12-06 15:50:08.866052525 +0000 UTC m=+248.756932111,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 15:50:16 crc kubenswrapper[4813]: I1206 15:50:16.490553 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:50:16 crc kubenswrapper[4813]: I1206 15:50:16.492359 4813 status_manager.go:851] "Failed to get status for pod" podUID="6a553ef5-b76a-46cd-9e49-30bf4ca86e29" pod="openshift-marketplace/redhat-operators-5zp6r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-5zp6r\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:16 crc kubenswrapper[4813]: I1206 15:50:16.492723 4813 status_manager.go:851] "Failed to get status for pod" podUID="952e1f68-97e3-4191-9dde-43d67e83a1e0" pod="openshift-marketplace/redhat-operators-7d6jw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-7d6jw\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:16 crc kubenswrapper[4813]: I1206 15:50:16.492960 4813 status_manager.go:851] "Failed to get status for pod" podUID="6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:16 crc kubenswrapper[4813]: I1206 15:50:16.506799 4813 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="378199a2-e738-42f0-9a44-a5722a68076a" Dec 06 15:50:16 crc kubenswrapper[4813]: I1206 15:50:16.506834 4813 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="378199a2-e738-42f0-9a44-a5722a68076a" Dec 06 15:50:16 crc kubenswrapper[4813]: E1206 15:50:16.507361 4813 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:50:16 crc kubenswrapper[4813]: I1206 15:50:16.507836 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:50:16 crc kubenswrapper[4813]: E1206 15:50:16.605057 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="6.4s" Dec 06 15:50:16 crc kubenswrapper[4813]: I1206 15:50:16.967876 4813 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="a88f856983312ebdc68361febe64cb69272536abd20f4e33c181eb8c32062fbb" exitCode=0 Dec 06 15:50:16 crc kubenswrapper[4813]: I1206 15:50:16.967988 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"a88f856983312ebdc68361febe64cb69272536abd20f4e33c181eb8c32062fbb"} Dec 06 15:50:16 crc kubenswrapper[4813]: I1206 15:50:16.968113 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"bef4414eff2edd95567754ee8742d50ffd649623084c85f5de2d071b132520b1"} Dec 06 15:50:16 crc kubenswrapper[4813]: I1206 15:50:16.968351 4813 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="378199a2-e738-42f0-9a44-a5722a68076a" Dec 06 15:50:16 crc kubenswrapper[4813]: I1206 15:50:16.968362 4813 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="378199a2-e738-42f0-9a44-a5722a68076a" Dec 06 15:50:16 crc kubenswrapper[4813]: I1206 15:50:16.968845 4813 status_manager.go:851] "Failed to get status for pod" podUID="6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:16 crc kubenswrapper[4813]: E1206 15:50:16.968878 4813 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:50:16 crc kubenswrapper[4813]: I1206 15:50:16.969245 4813 status_manager.go:851] "Failed to get status for pod" podUID="6a553ef5-b76a-46cd-9e49-30bf4ca86e29" pod="openshift-marketplace/redhat-operators-5zp6r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-5zp6r\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:16 crc kubenswrapper[4813]: I1206 15:50:16.969611 4813 status_manager.go:851] "Failed to get status for pod" podUID="952e1f68-97e3-4191-9dde-43d67e83a1e0" pod="openshift-marketplace/redhat-operators-7d6jw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-7d6jw\": dial tcp 38.102.83.9:6443: connect: connection refused" Dec 06 15:50:17 crc kubenswrapper[4813]: I1206 15:50:17.974903 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4e19b815576a2a911274fccba068c47ba181914bae9350450ac1295f6d4d4f75"} Dec 06 15:50:17 crc kubenswrapper[4813]: I1206 15:50:17.974950 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"13162267a84a697dbea55d9383ea75607869e8d5274bce767029ad4e9d26abfc"} Dec 06 15:50:17 crc kubenswrapper[4813]: I1206 15:50:17.974963 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d84301b526db03583751f04b6533db19b00bc1c23f5bb7fe97ef01f518e45157"} Dec 06 15:50:17 crc kubenswrapper[4813]: I1206 15:50:17.974975 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e37c301fcfd9c5493f3d10e6e5084f6baae576d5d4ac90daa8aac21039c4b598"} Dec 06 15:50:17 crc kubenswrapper[4813]: I1206 15:50:17.977511 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 06 15:50:17 crc kubenswrapper[4813]: I1206 15:50:17.977550 4813 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545" exitCode=1 Dec 06 15:50:17 crc kubenswrapper[4813]: I1206 15:50:17.977577 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545"} Dec 06 15:50:17 crc kubenswrapper[4813]: I1206 15:50:17.977984 4813 scope.go:117] "RemoveContainer" containerID="cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545" Dec 06 15:50:18 crc kubenswrapper[4813]: I1206 15:50:18.987493 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"fc4e2ac8d83d0d7e5d76286a38a6f8b3d9f941584144b7817706ae8d636969d3"} Dec 06 15:50:18 crc kubenswrapper[4813]: I1206 15:50:18.987662 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:50:18 crc kubenswrapper[4813]: I1206 15:50:18.987768 4813 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="378199a2-e738-42f0-9a44-a5722a68076a" Dec 06 15:50:18 crc kubenswrapper[4813]: I1206 15:50:18.987790 4813 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="378199a2-e738-42f0-9a44-a5722a68076a" Dec 06 15:50:18 crc kubenswrapper[4813]: I1206 15:50:18.990719 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 06 15:50:18 crc kubenswrapper[4813]: I1206 15:50:18.990767 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"17af73d1dbe8321c83f2bb87f0aff0cb05824a6c180b77b4869bb4fa29cc81b3"} Dec 06 15:50:21 crc kubenswrapper[4813]: I1206 15:50:21.509400 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:50:21 crc kubenswrapper[4813]: I1206 15:50:21.509737 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:50:21 crc kubenswrapper[4813]: I1206 15:50:21.515421 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:50:21 crc kubenswrapper[4813]: I1206 15:50:21.595153 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 15:50:21 crc kubenswrapper[4813]: I1206 15:50:21.596209 4813 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 06 15:50:21 crc kubenswrapper[4813]: I1206 15:50:21.596251 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 06 15:50:22 crc kubenswrapper[4813]: I1206 15:50:22.272008 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 15:50:23 crc kubenswrapper[4813]: I1206 15:50:23.995352 4813 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:50:24 crc kubenswrapper[4813]: I1206 15:50:24.013161 4813 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="0cbcefb0-9e50-4ac9-812e-8775b521727e" Dec 06 15:50:24 crc kubenswrapper[4813]: I1206 15:50:24.029965 4813 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="378199a2-e738-42f0-9a44-a5722a68076a" Dec 06 15:50:24 crc kubenswrapper[4813]: I1206 15:50:24.029994 4813 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="378199a2-e738-42f0-9a44-a5722a68076a" Dec 06 15:50:24 crc kubenswrapper[4813]: I1206 15:50:24.039427 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:50:24 crc kubenswrapper[4813]: I1206 15:50:24.040351 4813 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="0cbcefb0-9e50-4ac9-812e-8775b521727e" Dec 06 15:50:25 crc kubenswrapper[4813]: I1206 15:50:25.033664 4813 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="378199a2-e738-42f0-9a44-a5722a68076a" Dec 06 15:50:25 crc kubenswrapper[4813]: I1206 15:50:25.033875 4813 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="378199a2-e738-42f0-9a44-a5722a68076a" Dec 06 15:50:25 crc kubenswrapper[4813]: I1206 15:50:25.035897 4813 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="0cbcefb0-9e50-4ac9-812e-8775b521727e" Dec 06 15:50:31 crc kubenswrapper[4813]: I1206 15:50:31.594990 4813 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 06 15:50:31 crc kubenswrapper[4813]: I1206 15:50:31.595685 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 06 15:50:35 crc kubenswrapper[4813]: I1206 15:50:35.063963 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 06 15:50:35 crc kubenswrapper[4813]: I1206 15:50:35.512664 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 06 15:50:35 crc kubenswrapper[4813]: I1206 15:50:35.606193 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 06 15:50:35 crc kubenswrapper[4813]: I1206 15:50:35.621582 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 06 15:50:35 crc kubenswrapper[4813]: I1206 15:50:35.867895 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 06 15:50:35 crc kubenswrapper[4813]: I1206 15:50:35.934989 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 06 15:50:36 crc kubenswrapper[4813]: I1206 15:50:36.316712 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 06 15:50:36 crc kubenswrapper[4813]: I1206 15:50:36.518867 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 06 15:50:36 crc kubenswrapper[4813]: I1206 15:50:36.533294 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 06 15:50:36 crc kubenswrapper[4813]: I1206 15:50:36.538629 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 06 15:50:36 crc kubenswrapper[4813]: I1206 15:50:36.625222 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 06 15:50:36 crc kubenswrapper[4813]: I1206 15:50:36.704099 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 06 15:50:36 crc kubenswrapper[4813]: I1206 15:50:36.746254 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 06 15:50:36 crc kubenswrapper[4813]: I1206 15:50:36.816690 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 06 15:50:36 crc kubenswrapper[4813]: I1206 15:50:36.946903 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 06 15:50:37 crc kubenswrapper[4813]: I1206 15:50:37.241053 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 06 15:50:37 crc kubenswrapper[4813]: I1206 15:50:37.279641 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 06 15:50:37 crc kubenswrapper[4813]: I1206 15:50:37.286572 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 06 15:50:37 crc kubenswrapper[4813]: I1206 15:50:37.490151 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 06 15:50:37 crc kubenswrapper[4813]: I1206 15:50:37.492272 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 06 15:50:37 crc kubenswrapper[4813]: I1206 15:50:37.494866 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 06 15:50:37 crc kubenswrapper[4813]: I1206 15:50:37.505347 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 06 15:50:37 crc kubenswrapper[4813]: I1206 15:50:37.590135 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 06 15:50:37 crc kubenswrapper[4813]: I1206 15:50:37.595937 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 06 15:50:37 crc kubenswrapper[4813]: I1206 15:50:37.745958 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.067223 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.108797 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.136462 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.171918 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.203122 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.243500 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.320240 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.336959 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.349143 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.355401 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.383163 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.386630 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.396125 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.440737 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.493014 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.641932 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.643612 4813 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.652676 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.652878 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.653130 4813 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="378199a2-e738-42f0-9a44-a5722a68076a" Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.653152 4813 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="378199a2-e738-42f0-9a44-a5722a68076a" Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.659710 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.670520 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.687569 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=14.68754843 podStartE2EDuration="14.68754843s" podCreationTimestamp="2025-12-06 15:50:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:50:38.686671736 +0000 UTC m=+278.577551342" watchObservedRunningTime="2025-12-06 15:50:38.68754843 +0000 UTC m=+278.578428056" Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.710871 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.746976 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.758306 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.784605 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.800321 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.801669 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.846991 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.904882 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 06 15:50:38 crc kubenswrapper[4813]: I1206 15:50:38.905045 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 06 15:50:39 crc kubenswrapper[4813]: I1206 15:50:39.115105 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 06 15:50:39 crc kubenswrapper[4813]: I1206 15:50:39.165388 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 06 15:50:39 crc kubenswrapper[4813]: I1206 15:50:39.223874 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 06 15:50:39 crc kubenswrapper[4813]: I1206 15:50:39.235472 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 06 15:50:39 crc kubenswrapper[4813]: I1206 15:50:39.348343 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 06 15:50:39 crc kubenswrapper[4813]: I1206 15:50:39.364140 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 06 15:50:39 crc kubenswrapper[4813]: I1206 15:50:39.373224 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 06 15:50:39 crc kubenswrapper[4813]: I1206 15:50:39.423222 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 06 15:50:39 crc kubenswrapper[4813]: I1206 15:50:39.487835 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 06 15:50:39 crc kubenswrapper[4813]: I1206 15:50:39.515690 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 06 15:50:39 crc kubenswrapper[4813]: I1206 15:50:39.525381 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 06 15:50:39 crc kubenswrapper[4813]: I1206 15:50:39.538866 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 06 15:50:39 crc kubenswrapper[4813]: I1206 15:50:39.620838 4813 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 06 15:50:39 crc kubenswrapper[4813]: I1206 15:50:39.686800 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 06 15:50:39 crc kubenswrapper[4813]: I1206 15:50:39.698734 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 06 15:50:39 crc kubenswrapper[4813]: I1206 15:50:39.701189 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 06 15:50:39 crc kubenswrapper[4813]: I1206 15:50:39.819766 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 06 15:50:39 crc kubenswrapper[4813]: I1206 15:50:39.848083 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 06 15:50:39 crc kubenswrapper[4813]: I1206 15:50:39.860334 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 06 15:50:39 crc kubenswrapper[4813]: I1206 15:50:39.986803 4813 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 06 15:50:40 crc kubenswrapper[4813]: I1206 15:50:40.038685 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 06 15:50:40 crc kubenswrapper[4813]: I1206 15:50:40.095875 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 06 15:50:40 crc kubenswrapper[4813]: I1206 15:50:40.133193 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 06 15:50:40 crc kubenswrapper[4813]: I1206 15:50:40.216035 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 06 15:50:40 crc kubenswrapper[4813]: I1206 15:50:40.301461 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 06 15:50:40 crc kubenswrapper[4813]: I1206 15:50:40.342751 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 06 15:50:40 crc kubenswrapper[4813]: I1206 15:50:40.420091 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 06 15:50:40 crc kubenswrapper[4813]: I1206 15:50:40.491856 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 06 15:50:40 crc kubenswrapper[4813]: I1206 15:50:40.523539 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 06 15:50:40 crc kubenswrapper[4813]: I1206 15:50:40.654393 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 06 15:50:40 crc kubenswrapper[4813]: I1206 15:50:40.697984 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 06 15:50:40 crc kubenswrapper[4813]: I1206 15:50:40.714240 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 06 15:50:40 crc kubenswrapper[4813]: I1206 15:50:40.954597 4813 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 06 15:50:40 crc kubenswrapper[4813]: I1206 15:50:40.970693 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 06 15:50:41 crc kubenswrapper[4813]: I1206 15:50:41.056228 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 06 15:50:41 crc kubenswrapper[4813]: I1206 15:50:41.057448 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 06 15:50:41 crc kubenswrapper[4813]: I1206 15:50:41.143615 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 06 15:50:41 crc kubenswrapper[4813]: I1206 15:50:41.173842 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 06 15:50:41 crc kubenswrapper[4813]: I1206 15:50:41.210302 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 06 15:50:41 crc kubenswrapper[4813]: I1206 15:50:41.266214 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 06 15:50:41 crc kubenswrapper[4813]: I1206 15:50:41.269889 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 06 15:50:41 crc kubenswrapper[4813]: I1206 15:50:41.460522 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 06 15:50:41 crc kubenswrapper[4813]: I1206 15:50:41.465994 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 06 15:50:41 crc kubenswrapper[4813]: I1206 15:50:41.527469 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qnq74"] Dec 06 15:50:41 crc kubenswrapper[4813]: I1206 15:50:41.527858 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qnq74" podUID="f522807f-660d-4791-95f4-f2d9daa58936" containerName="registry-server" containerID="cri-o://ab86ee25c416c62178029b32a6c3639723f9b985501d0ac91049782cf3d230be" gracePeriod=2 Dec 06 15:50:41 crc kubenswrapper[4813]: I1206 15:50:41.554909 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 06 15:50:41 crc kubenswrapper[4813]: I1206 15:50:41.595792 4813 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 06 15:50:41 crc kubenswrapper[4813]: I1206 15:50:41.595855 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 06 15:50:41 crc kubenswrapper[4813]: I1206 15:50:41.595905 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 15:50:41 crc kubenswrapper[4813]: I1206 15:50:41.596773 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"17af73d1dbe8321c83f2bb87f0aff0cb05824a6c180b77b4869bb4fa29cc81b3"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Dec 06 15:50:41 crc kubenswrapper[4813]: I1206 15:50:41.596896 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://17af73d1dbe8321c83f2bb87f0aff0cb05824a6c180b77b4869bb4fa29cc81b3" gracePeriod=30 Dec 06 15:50:41 crc kubenswrapper[4813]: I1206 15:50:41.645798 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 06 15:50:41 crc kubenswrapper[4813]: I1206 15:50:41.662769 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 06 15:50:41 crc kubenswrapper[4813]: I1206 15:50:41.785965 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 06 15:50:41 crc kubenswrapper[4813]: I1206 15:50:41.875415 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 06 15:50:41 crc kubenswrapper[4813]: I1206 15:50:41.878271 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 06 15:50:41 crc kubenswrapper[4813]: I1206 15:50:41.922840 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 06 15:50:41 crc kubenswrapper[4813]: I1206 15:50:41.930066 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qnq74" Dec 06 15:50:41 crc kubenswrapper[4813]: I1206 15:50:41.985471 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.018652 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.018700 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.023225 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.030974 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.037019 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.044860 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.059540 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.091569 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.100687 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59zbh\" (UniqueName: \"kubernetes.io/projected/f522807f-660d-4791-95f4-f2d9daa58936-kube-api-access-59zbh\") pod \"f522807f-660d-4791-95f4-f2d9daa58936\" (UID: \"f522807f-660d-4791-95f4-f2d9daa58936\") " Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.100729 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f522807f-660d-4791-95f4-f2d9daa58936-utilities\") pod \"f522807f-660d-4791-95f4-f2d9daa58936\" (UID: \"f522807f-660d-4791-95f4-f2d9daa58936\") " Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.100754 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f522807f-660d-4791-95f4-f2d9daa58936-catalog-content\") pod \"f522807f-660d-4791-95f4-f2d9daa58936\" (UID: \"f522807f-660d-4791-95f4-f2d9daa58936\") " Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.102208 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f522807f-660d-4791-95f4-f2d9daa58936-utilities" (OuterVolumeSpecName: "utilities") pod "f522807f-660d-4791-95f4-f2d9daa58936" (UID: "f522807f-660d-4791-95f4-f2d9daa58936"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.113474 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f522807f-660d-4791-95f4-f2d9daa58936-kube-api-access-59zbh" (OuterVolumeSpecName: "kube-api-access-59zbh") pod "f522807f-660d-4791-95f4-f2d9daa58936" (UID: "f522807f-660d-4791-95f4-f2d9daa58936"). InnerVolumeSpecName "kube-api-access-59zbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.115185 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.141944 4813 generic.go:334] "Generic (PLEG): container finished" podID="f522807f-660d-4791-95f4-f2d9daa58936" containerID="ab86ee25c416c62178029b32a6c3639723f9b985501d0ac91049782cf3d230be" exitCode=0 Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.141986 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qnq74" event={"ID":"f522807f-660d-4791-95f4-f2d9daa58936","Type":"ContainerDied","Data":"ab86ee25c416c62178029b32a6c3639723f9b985501d0ac91049782cf3d230be"} Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.142023 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qnq74" event={"ID":"f522807f-660d-4791-95f4-f2d9daa58936","Type":"ContainerDied","Data":"c1fe3e4f52a8a698a90dfd2bff093f40e85bc638cf468d8f9f9e8f43f475a0f2"} Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.142030 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qnq74" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.142048 4813 scope.go:117] "RemoveContainer" containerID="ab86ee25c416c62178029b32a6c3639723f9b985501d0ac91049782cf3d230be" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.158912 4813 scope.go:117] "RemoveContainer" containerID="98c3136dd68e10777f5a954d2dc9f8a9e069ef68281dc21b5388a3a6125bf467" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.164665 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f522807f-660d-4791-95f4-f2d9daa58936-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f522807f-660d-4791-95f4-f2d9daa58936" (UID: "f522807f-660d-4791-95f4-f2d9daa58936"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.178066 4813 scope.go:117] "RemoveContainer" containerID="ad14489bc0eea912c65421f84a92ce0d9ae9028f677db47ed06e5142e4929884" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.193407 4813 scope.go:117] "RemoveContainer" containerID="ab86ee25c416c62178029b32a6c3639723f9b985501d0ac91049782cf3d230be" Dec 06 15:50:42 crc kubenswrapper[4813]: E1206 15:50:42.193774 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab86ee25c416c62178029b32a6c3639723f9b985501d0ac91049782cf3d230be\": container with ID starting with ab86ee25c416c62178029b32a6c3639723f9b985501d0ac91049782cf3d230be not found: ID does not exist" containerID="ab86ee25c416c62178029b32a6c3639723f9b985501d0ac91049782cf3d230be" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.193807 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab86ee25c416c62178029b32a6c3639723f9b985501d0ac91049782cf3d230be"} err="failed to get container status \"ab86ee25c416c62178029b32a6c3639723f9b985501d0ac91049782cf3d230be\": rpc error: code = NotFound desc = could not find container \"ab86ee25c416c62178029b32a6c3639723f9b985501d0ac91049782cf3d230be\": container with ID starting with ab86ee25c416c62178029b32a6c3639723f9b985501d0ac91049782cf3d230be not found: ID does not exist" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.193829 4813 scope.go:117] "RemoveContainer" containerID="98c3136dd68e10777f5a954d2dc9f8a9e069ef68281dc21b5388a3a6125bf467" Dec 06 15:50:42 crc kubenswrapper[4813]: E1206 15:50:42.194002 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98c3136dd68e10777f5a954d2dc9f8a9e069ef68281dc21b5388a3a6125bf467\": container with ID starting with 98c3136dd68e10777f5a954d2dc9f8a9e069ef68281dc21b5388a3a6125bf467 not found: ID does not exist" containerID="98c3136dd68e10777f5a954d2dc9f8a9e069ef68281dc21b5388a3a6125bf467" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.194034 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98c3136dd68e10777f5a954d2dc9f8a9e069ef68281dc21b5388a3a6125bf467"} err="failed to get container status \"98c3136dd68e10777f5a954d2dc9f8a9e069ef68281dc21b5388a3a6125bf467\": rpc error: code = NotFound desc = could not find container \"98c3136dd68e10777f5a954d2dc9f8a9e069ef68281dc21b5388a3a6125bf467\": container with ID starting with 98c3136dd68e10777f5a954d2dc9f8a9e069ef68281dc21b5388a3a6125bf467 not found: ID does not exist" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.194053 4813 scope.go:117] "RemoveContainer" containerID="ad14489bc0eea912c65421f84a92ce0d9ae9028f677db47ed06e5142e4929884" Dec 06 15:50:42 crc kubenswrapper[4813]: E1206 15:50:42.194591 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad14489bc0eea912c65421f84a92ce0d9ae9028f677db47ed06e5142e4929884\": container with ID starting with ad14489bc0eea912c65421f84a92ce0d9ae9028f677db47ed06e5142e4929884 not found: ID does not exist" containerID="ad14489bc0eea912c65421f84a92ce0d9ae9028f677db47ed06e5142e4929884" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.194641 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad14489bc0eea912c65421f84a92ce0d9ae9028f677db47ed06e5142e4929884"} err="failed to get container status \"ad14489bc0eea912c65421f84a92ce0d9ae9028f677db47ed06e5142e4929884\": rpc error: code = NotFound desc = could not find container \"ad14489bc0eea912c65421f84a92ce0d9ae9028f677db47ed06e5142e4929884\": container with ID starting with ad14489bc0eea912c65421f84a92ce0d9ae9028f677db47ed06e5142e4929884 not found: ID does not exist" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.195707 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.202496 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59zbh\" (UniqueName: \"kubernetes.io/projected/f522807f-660d-4791-95f4-f2d9daa58936-kube-api-access-59zbh\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.202520 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f522807f-660d-4791-95f4-f2d9daa58936-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.202531 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f522807f-660d-4791-95f4-f2d9daa58936-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.207542 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.238712 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.245981 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.269818 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.324576 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.353303 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.372911 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.437222 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.475863 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.495735 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qnq74"] Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.499429 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qnq74"] Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.516809 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.518351 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.535689 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.570868 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.640683 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.672489 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.696465 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.700691 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 06 15:50:42 crc kubenswrapper[4813]: I1206 15:50:42.901765 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 06 15:50:43 crc kubenswrapper[4813]: I1206 15:50:43.066911 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 06 15:50:43 crc kubenswrapper[4813]: I1206 15:50:43.108706 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 06 15:50:43 crc kubenswrapper[4813]: I1206 15:50:43.135677 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 06 15:50:43 crc kubenswrapper[4813]: I1206 15:50:43.267096 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 06 15:50:43 crc kubenswrapper[4813]: I1206 15:50:43.328098 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 06 15:50:43 crc kubenswrapper[4813]: I1206 15:50:43.385027 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 06 15:50:43 crc kubenswrapper[4813]: I1206 15:50:43.390105 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 06 15:50:43 crc kubenswrapper[4813]: I1206 15:50:43.412085 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 06 15:50:43 crc kubenswrapper[4813]: I1206 15:50:43.512047 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 06 15:50:43 crc kubenswrapper[4813]: I1206 15:50:43.512473 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 06 15:50:43 crc kubenswrapper[4813]: I1206 15:50:43.522559 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 06 15:50:43 crc kubenswrapper[4813]: I1206 15:50:43.653457 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 06 15:50:43 crc kubenswrapper[4813]: I1206 15:50:43.679443 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 06 15:50:43 crc kubenswrapper[4813]: I1206 15:50:43.762886 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 06 15:50:43 crc kubenswrapper[4813]: I1206 15:50:43.872394 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 06 15:50:44 crc kubenswrapper[4813]: I1206 15:50:44.109011 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 06 15:50:44 crc kubenswrapper[4813]: I1206 15:50:44.158414 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 06 15:50:44 crc kubenswrapper[4813]: I1206 15:50:44.232785 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 06 15:50:44 crc kubenswrapper[4813]: I1206 15:50:44.300796 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 06 15:50:44 crc kubenswrapper[4813]: I1206 15:50:44.308762 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 06 15:50:44 crc kubenswrapper[4813]: I1206 15:50:44.347246 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 06 15:50:44 crc kubenswrapper[4813]: I1206 15:50:44.371051 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 06 15:50:44 crc kubenswrapper[4813]: I1206 15:50:44.419438 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 06 15:50:44 crc kubenswrapper[4813]: I1206 15:50:44.506391 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f522807f-660d-4791-95f4-f2d9daa58936" path="/var/lib/kubelet/pods/f522807f-660d-4791-95f4-f2d9daa58936/volumes" Dec 06 15:50:44 crc kubenswrapper[4813]: I1206 15:50:44.566908 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 06 15:50:44 crc kubenswrapper[4813]: I1206 15:50:44.622925 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 06 15:50:44 crc kubenswrapper[4813]: I1206 15:50:44.669399 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 06 15:50:44 crc kubenswrapper[4813]: I1206 15:50:44.718467 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 06 15:50:44 crc kubenswrapper[4813]: I1206 15:50:44.811372 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 06 15:50:44 crc kubenswrapper[4813]: I1206 15:50:44.812039 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 06 15:50:44 crc kubenswrapper[4813]: I1206 15:50:44.816168 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 06 15:50:44 crc kubenswrapper[4813]: I1206 15:50:44.844827 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 06 15:50:44 crc kubenswrapper[4813]: I1206 15:50:44.910319 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 06 15:50:45 crc kubenswrapper[4813]: I1206 15:50:45.002576 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 06 15:50:45 crc kubenswrapper[4813]: I1206 15:50:45.039580 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 06 15:50:45 crc kubenswrapper[4813]: I1206 15:50:45.057393 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 06 15:50:45 crc kubenswrapper[4813]: I1206 15:50:45.060650 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 06 15:50:45 crc kubenswrapper[4813]: I1206 15:50:45.071156 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 06 15:50:45 crc kubenswrapper[4813]: I1206 15:50:45.099007 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 06 15:50:45 crc kubenswrapper[4813]: I1206 15:50:45.172809 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 06 15:50:45 crc kubenswrapper[4813]: I1206 15:50:45.262500 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 06 15:50:45 crc kubenswrapper[4813]: I1206 15:50:45.424467 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 06 15:50:45 crc kubenswrapper[4813]: I1206 15:50:45.464036 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 06 15:50:45 crc kubenswrapper[4813]: I1206 15:50:45.562911 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 06 15:50:45 crc kubenswrapper[4813]: I1206 15:50:45.569816 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 06 15:50:45 crc kubenswrapper[4813]: I1206 15:50:45.577031 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 06 15:50:45 crc kubenswrapper[4813]: I1206 15:50:45.699719 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 06 15:50:45 crc kubenswrapper[4813]: I1206 15:50:45.723551 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 06 15:50:45 crc kubenswrapper[4813]: I1206 15:50:45.730466 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 06 15:50:45 crc kubenswrapper[4813]: I1206 15:50:45.784397 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 06 15:50:45 crc kubenswrapper[4813]: I1206 15:50:45.865051 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 06 15:50:45 crc kubenswrapper[4813]: I1206 15:50:45.964681 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 06 15:50:46 crc kubenswrapper[4813]: I1206 15:50:46.046176 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 06 15:50:46 crc kubenswrapper[4813]: I1206 15:50:46.067360 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 06 15:50:46 crc kubenswrapper[4813]: I1206 15:50:46.182557 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 06 15:50:46 crc kubenswrapper[4813]: I1206 15:50:46.518153 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 06 15:50:46 crc kubenswrapper[4813]: I1206 15:50:46.556864 4813 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 06 15:50:46 crc kubenswrapper[4813]: I1206 15:50:46.613559 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 06 15:50:46 crc kubenswrapper[4813]: I1206 15:50:46.642364 4813 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 06 15:50:46 crc kubenswrapper[4813]: I1206 15:50:46.642715 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://0eb54412ce9438f87b8eb1721001f4d5bb20f02f364dc78f126722417786f8ac" gracePeriod=5 Dec 06 15:50:46 crc kubenswrapper[4813]: I1206 15:50:46.694236 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 06 15:50:46 crc kubenswrapper[4813]: I1206 15:50:46.782464 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 06 15:50:46 crc kubenswrapper[4813]: I1206 15:50:46.851098 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 06 15:50:46 crc kubenswrapper[4813]: I1206 15:50:46.861467 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 06 15:50:46 crc kubenswrapper[4813]: I1206 15:50:46.911449 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 06 15:50:46 crc kubenswrapper[4813]: I1206 15:50:46.986505 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 06 15:50:47 crc kubenswrapper[4813]: I1206 15:50:47.097089 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 06 15:50:47 crc kubenswrapper[4813]: I1206 15:50:47.187142 4813 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 06 15:50:47 crc kubenswrapper[4813]: I1206 15:50:47.380502 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 06 15:50:47 crc kubenswrapper[4813]: I1206 15:50:47.412304 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 06 15:50:47 crc kubenswrapper[4813]: I1206 15:50:47.430035 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 06 15:50:47 crc kubenswrapper[4813]: I1206 15:50:47.461254 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 06 15:50:47 crc kubenswrapper[4813]: I1206 15:50:47.473210 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 06 15:50:47 crc kubenswrapper[4813]: I1206 15:50:47.527396 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 06 15:50:47 crc kubenswrapper[4813]: I1206 15:50:47.561387 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 06 15:50:47 crc kubenswrapper[4813]: I1206 15:50:47.597056 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 06 15:50:47 crc kubenswrapper[4813]: I1206 15:50:47.602020 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 06 15:50:47 crc kubenswrapper[4813]: I1206 15:50:47.625153 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 06 15:50:47 crc kubenswrapper[4813]: I1206 15:50:47.732731 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 06 15:50:47 crc kubenswrapper[4813]: I1206 15:50:47.857492 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 06 15:50:47 crc kubenswrapper[4813]: I1206 15:50:47.886826 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 06 15:50:47 crc kubenswrapper[4813]: I1206 15:50:47.991322 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 06 15:50:48 crc kubenswrapper[4813]: I1206 15:50:48.064687 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 06 15:50:48 crc kubenswrapper[4813]: I1206 15:50:48.100195 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 06 15:50:48 crc kubenswrapper[4813]: I1206 15:50:48.108229 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 06 15:50:48 crc kubenswrapper[4813]: I1206 15:50:48.161443 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 06 15:50:48 crc kubenswrapper[4813]: I1206 15:50:48.166817 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 06 15:50:48 crc kubenswrapper[4813]: I1206 15:50:48.234343 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 06 15:50:48 crc kubenswrapper[4813]: I1206 15:50:48.263539 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 06 15:50:48 crc kubenswrapper[4813]: I1206 15:50:48.274292 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 06 15:50:48 crc kubenswrapper[4813]: I1206 15:50:48.314186 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 06 15:50:48 crc kubenswrapper[4813]: I1206 15:50:48.431983 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 06 15:50:48 crc kubenswrapper[4813]: I1206 15:50:48.506692 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 06 15:50:48 crc kubenswrapper[4813]: I1206 15:50:48.523222 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 06 15:50:48 crc kubenswrapper[4813]: I1206 15:50:48.585305 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 06 15:50:48 crc kubenswrapper[4813]: I1206 15:50:48.649954 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 06 15:50:48 crc kubenswrapper[4813]: I1206 15:50:48.690537 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 06 15:50:48 crc kubenswrapper[4813]: I1206 15:50:48.715390 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 06 15:50:48 crc kubenswrapper[4813]: I1206 15:50:48.823946 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 06 15:50:48 crc kubenswrapper[4813]: I1206 15:50:48.869802 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 06 15:50:48 crc kubenswrapper[4813]: I1206 15:50:48.917986 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 06 15:50:48 crc kubenswrapper[4813]: I1206 15:50:48.984709 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 06 15:50:49 crc kubenswrapper[4813]: I1206 15:50:49.115988 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 06 15:50:49 crc kubenswrapper[4813]: I1206 15:50:49.162822 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 06 15:50:49 crc kubenswrapper[4813]: I1206 15:50:49.296916 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 06 15:50:49 crc kubenswrapper[4813]: I1206 15:50:49.319127 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 06 15:50:49 crc kubenswrapper[4813]: I1206 15:50:49.334616 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 06 15:50:49 crc kubenswrapper[4813]: I1206 15:50:49.385849 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 06 15:50:49 crc kubenswrapper[4813]: I1206 15:50:49.490222 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 06 15:50:49 crc kubenswrapper[4813]: I1206 15:50:49.630950 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 06 15:50:49 crc kubenswrapper[4813]: I1206 15:50:49.824224 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 06 15:50:49 crc kubenswrapper[4813]: I1206 15:50:49.829351 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 06 15:50:49 crc kubenswrapper[4813]: I1206 15:50:49.889013 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 06 15:50:49 crc kubenswrapper[4813]: I1206 15:50:49.896254 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 06 15:50:49 crc kubenswrapper[4813]: I1206 15:50:49.953868 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 06 15:50:50 crc kubenswrapper[4813]: I1206 15:50:50.314149 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 06 15:50:50 crc kubenswrapper[4813]: I1206 15:50:50.322507 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 06 15:50:50 crc kubenswrapper[4813]: I1206 15:50:50.531380 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-54xzm"] Dec 06 15:50:50 crc kubenswrapper[4813]: I1206 15:50:50.531957 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-54xzm" podUID="e2fac5b3-c60b-4c96-8841-ae53a24aaed1" containerName="registry-server" containerID="cri-o://58ab538dbec20a657dfd0e11d156896a425cf1cf03ac50cbbb65c4454607e6f2" gracePeriod=30 Dec 06 15:50:50 crc kubenswrapper[4813]: I1206 15:50:50.535004 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kp6t2"] Dec 06 15:50:50 crc kubenswrapper[4813]: I1206 15:50:50.547165 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6rwpr"] Dec 06 15:50:50 crc kubenswrapper[4813]: I1206 15:50:50.547496 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6rwpr" podUID="76073738-387a-4968-a483-8880e866d1e4" containerName="registry-server" containerID="cri-o://dc210c0684fff6a7e1dbf1d2841de30e46e582fb956d958e6cb6ed48c0837da8" gracePeriod=30 Dec 06 15:50:50 crc kubenswrapper[4813]: I1206 15:50:50.547803 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kp6t2" podUID="e982bb16-5f82-418b-8c03-8c73d22e010b" containerName="registry-server" containerID="cri-o://6e4cbadc5a42dd6f512afa67014f67d6c1c96d3dbe1b330a18894f6c5d98b8d3" gracePeriod=30 Dec 06 15:50:50 crc kubenswrapper[4813]: I1206 15:50:50.574102 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-289lx"] Dec 06 15:50:50 crc kubenswrapper[4813]: I1206 15:50:50.574392 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-289lx" podUID="60b3916a-79fc-42b6-86c6-4cae051457ca" containerName="marketplace-operator" containerID="cri-o://441756359cdf72b3e97cab1b8ce564665e57f14ec2279db9364d1706faa70be5" gracePeriod=30 Dec 06 15:50:50 crc kubenswrapper[4813]: I1206 15:50:50.582979 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vn7fl"] Dec 06 15:50:50 crc kubenswrapper[4813]: I1206 15:50:50.583253 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vn7fl" podUID="da4de6a4-9efc-425c-8712-e66919babdf2" containerName="registry-server" containerID="cri-o://6041b776a12e9bbefe8831d00231faeec17ad0e52b7045f06b8132aa5d850c9c" gracePeriod=30 Dec 06 15:50:50 crc kubenswrapper[4813]: I1206 15:50:50.587452 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5zp6r"] Dec 06 15:50:50 crc kubenswrapper[4813]: I1206 15:50:50.587797 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5zp6r" podUID="6a553ef5-b76a-46cd-9e49-30bf4ca86e29" containerName="registry-server" containerID="cri-o://5ff03ddb8621464af047b3a4ddbde8a63a7fd6cf544c0e8e25ca07066341a555" gracePeriod=30 Dec 06 15:50:50 crc kubenswrapper[4813]: I1206 15:50:50.596416 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7d6jw"] Dec 06 15:50:50 crc kubenswrapper[4813]: I1206 15:50:50.596621 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7d6jw" podUID="952e1f68-97e3-4191-9dde-43d67e83a1e0" containerName="registry-server" containerID="cri-o://fc5c2d08cc45d223b3dddb5282d706c77ec8854aa99c6a85001fec025351536f" gracePeriod=30 Dec 06 15:50:50 crc kubenswrapper[4813]: I1206 15:50:50.607638 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kp6t2"] Dec 06 15:50:50 crc kubenswrapper[4813]: I1206 15:50:50.685000 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 06 15:50:50 crc kubenswrapper[4813]: I1206 15:50:50.740818 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 06 15:50:50 crc kubenswrapper[4813]: E1206 15:50:50.909326 4813 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 58ab538dbec20a657dfd0e11d156896a425cf1cf03ac50cbbb65c4454607e6f2 is running failed: container process not found" containerID="58ab538dbec20a657dfd0e11d156896a425cf1cf03ac50cbbb65c4454607e6f2" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 15:50:50 crc kubenswrapper[4813]: E1206 15:50:50.909617 4813 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 58ab538dbec20a657dfd0e11d156896a425cf1cf03ac50cbbb65c4454607e6f2 is running failed: container process not found" containerID="58ab538dbec20a657dfd0e11d156896a425cf1cf03ac50cbbb65c4454607e6f2" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 15:50:50 crc kubenswrapper[4813]: E1206 15:50:50.909807 4813 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 58ab538dbec20a657dfd0e11d156896a425cf1cf03ac50cbbb65c4454607e6f2 is running failed: container process not found" containerID="58ab538dbec20a657dfd0e11d156896a425cf1cf03ac50cbbb65c4454607e6f2" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 15:50:50 crc kubenswrapper[4813]: E1206 15:50:50.909834 4813 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 58ab538dbec20a657dfd0e11d156896a425cf1cf03ac50cbbb65c4454607e6f2 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-54xzm" podUID="e2fac5b3-c60b-4c96-8841-ae53a24aaed1" containerName="registry-server" Dec 06 15:50:50 crc kubenswrapper[4813]: I1206 15:50:50.999068 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-54xzm" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.002636 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6rwpr" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.028835 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5zp6r" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.123102 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2fac5b3-c60b-4c96-8841-ae53a24aaed1-utilities\") pod \"e2fac5b3-c60b-4c96-8841-ae53a24aaed1\" (UID: \"e2fac5b3-c60b-4c96-8841-ae53a24aaed1\") " Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.123186 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqpfc\" (UniqueName: \"kubernetes.io/projected/76073738-387a-4968-a483-8880e866d1e4-kube-api-access-nqpfc\") pod \"76073738-387a-4968-a483-8880e866d1e4\" (UID: \"76073738-387a-4968-a483-8880e866d1e4\") " Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.123221 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76073738-387a-4968-a483-8880e866d1e4-utilities\") pod \"76073738-387a-4968-a483-8880e866d1e4\" (UID: \"76073738-387a-4968-a483-8880e866d1e4\") " Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.123315 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2fac5b3-c60b-4c96-8841-ae53a24aaed1-catalog-content\") pod \"e2fac5b3-c60b-4c96-8841-ae53a24aaed1\" (UID: \"e2fac5b3-c60b-4c96-8841-ae53a24aaed1\") " Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.123341 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k59xj\" (UniqueName: \"kubernetes.io/projected/e2fac5b3-c60b-4c96-8841-ae53a24aaed1-kube-api-access-k59xj\") pod \"e2fac5b3-c60b-4c96-8841-ae53a24aaed1\" (UID: \"e2fac5b3-c60b-4c96-8841-ae53a24aaed1\") " Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.123357 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76073738-387a-4968-a483-8880e866d1e4-catalog-content\") pod \"76073738-387a-4968-a483-8880e866d1e4\" (UID: \"76073738-387a-4968-a483-8880e866d1e4\") " Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.124346 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76073738-387a-4968-a483-8880e866d1e4-utilities" (OuterVolumeSpecName: "utilities") pod "76073738-387a-4968-a483-8880e866d1e4" (UID: "76073738-387a-4968-a483-8880e866d1e4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.124665 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2fac5b3-c60b-4c96-8841-ae53a24aaed1-utilities" (OuterVolumeSpecName: "utilities") pod "e2fac5b3-c60b-4c96-8841-ae53a24aaed1" (UID: "e2fac5b3-c60b-4c96-8841-ae53a24aaed1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.136085 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2fac5b3-c60b-4c96-8841-ae53a24aaed1-kube-api-access-k59xj" (OuterVolumeSpecName: "kube-api-access-k59xj") pod "e2fac5b3-c60b-4c96-8841-ae53a24aaed1" (UID: "e2fac5b3-c60b-4c96-8841-ae53a24aaed1"). InnerVolumeSpecName "kube-api-access-k59xj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.138543 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76073738-387a-4968-a483-8880e866d1e4-kube-api-access-nqpfc" (OuterVolumeSpecName: "kube-api-access-nqpfc") pod "76073738-387a-4968-a483-8880e866d1e4" (UID: "76073738-387a-4968-a483-8880e866d1e4"). InnerVolumeSpecName "kube-api-access-nqpfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.184126 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2fac5b3-c60b-4c96-8841-ae53a24aaed1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e2fac5b3-c60b-4c96-8841-ae53a24aaed1" (UID: "e2fac5b3-c60b-4c96-8841-ae53a24aaed1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.196555 4813 generic.go:334] "Generic (PLEG): container finished" podID="e2fac5b3-c60b-4c96-8841-ae53a24aaed1" containerID="58ab538dbec20a657dfd0e11d156896a425cf1cf03ac50cbbb65c4454607e6f2" exitCode=0 Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.196618 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-54xzm" event={"ID":"e2fac5b3-c60b-4c96-8841-ae53a24aaed1","Type":"ContainerDied","Data":"58ab538dbec20a657dfd0e11d156896a425cf1cf03ac50cbbb65c4454607e6f2"} Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.196666 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-54xzm" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.196681 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-54xzm" event={"ID":"e2fac5b3-c60b-4c96-8841-ae53a24aaed1","Type":"ContainerDied","Data":"bff1c33004538532a4b7d28d51ac7266dd442a1933568cb36ecb5430816e9a09"} Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.196695 4813 scope.go:117] "RemoveContainer" containerID="58ab538dbec20a657dfd0e11d156896a425cf1cf03ac50cbbb65c4454607e6f2" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.201643 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76073738-387a-4968-a483-8880e866d1e4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "76073738-387a-4968-a483-8880e866d1e4" (UID: "76073738-387a-4968-a483-8880e866d1e4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.202745 4813 generic.go:334] "Generic (PLEG): container finished" podID="952e1f68-97e3-4191-9dde-43d67e83a1e0" containerID="fc5c2d08cc45d223b3dddb5282d706c77ec8854aa99c6a85001fec025351536f" exitCode=0 Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.202814 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7d6jw" event={"ID":"952e1f68-97e3-4191-9dde-43d67e83a1e0","Type":"ContainerDied","Data":"fc5c2d08cc45d223b3dddb5282d706c77ec8854aa99c6a85001fec025351536f"} Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.207279 4813 generic.go:334] "Generic (PLEG): container finished" podID="e982bb16-5f82-418b-8c03-8c73d22e010b" containerID="6e4cbadc5a42dd6f512afa67014f67d6c1c96d3dbe1b330a18894f6c5d98b8d3" exitCode=0 Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.207342 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kp6t2" event={"ID":"e982bb16-5f82-418b-8c03-8c73d22e010b","Type":"ContainerDied","Data":"6e4cbadc5a42dd6f512afa67014f67d6c1c96d3dbe1b330a18894f6c5d98b8d3"} Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.210571 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vn7fl" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.210958 4813 generic.go:334] "Generic (PLEG): container finished" podID="60b3916a-79fc-42b6-86c6-4cae051457ca" containerID="441756359cdf72b3e97cab1b8ce564665e57f14ec2279db9364d1706faa70be5" exitCode=0 Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.210998 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-289lx" event={"ID":"60b3916a-79fc-42b6-86c6-4cae051457ca","Type":"ContainerDied","Data":"441756359cdf72b3e97cab1b8ce564665e57f14ec2279db9364d1706faa70be5"} Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.213034 4813 generic.go:334] "Generic (PLEG): container finished" podID="da4de6a4-9efc-425c-8712-e66919babdf2" containerID="6041b776a12e9bbefe8831d00231faeec17ad0e52b7045f06b8132aa5d850c9c" exitCode=0 Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.213175 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vn7fl" event={"ID":"da4de6a4-9efc-425c-8712-e66919babdf2","Type":"ContainerDied","Data":"6041b776a12e9bbefe8831d00231faeec17ad0e52b7045f06b8132aa5d850c9c"} Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.213205 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vn7fl" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.216631 4813 generic.go:334] "Generic (PLEG): container finished" podID="6a553ef5-b76a-46cd-9e49-30bf4ca86e29" containerID="5ff03ddb8621464af047b3a4ddbde8a63a7fd6cf544c0e8e25ca07066341a555" exitCode=0 Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.216724 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5zp6r" event={"ID":"6a553ef5-b76a-46cd-9e49-30bf4ca86e29","Type":"ContainerDied","Data":"5ff03ddb8621464af047b3a4ddbde8a63a7fd6cf544c0e8e25ca07066341a555"} Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.216766 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5zp6r" event={"ID":"6a553ef5-b76a-46cd-9e49-30bf4ca86e29","Type":"ContainerDied","Data":"9b87ae2c54a10257d7fd192ae57c948eaeadaeba991d0c9e46506b682085ece9"} Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.216842 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5zp6r" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.218788 4813 generic.go:334] "Generic (PLEG): container finished" podID="76073738-387a-4968-a483-8880e866d1e4" containerID="dc210c0684fff6a7e1dbf1d2841de30e46e582fb956d958e6cb6ed48c0837da8" exitCode=0 Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.218817 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rwpr" event={"ID":"76073738-387a-4968-a483-8880e866d1e4","Type":"ContainerDied","Data":"dc210c0684fff6a7e1dbf1d2841de30e46e582fb956d958e6cb6ed48c0837da8"} Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.218850 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6rwpr" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.218860 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rwpr" event={"ID":"76073738-387a-4968-a483-8880e866d1e4","Type":"ContainerDied","Data":"c3981bff3b38eea82b4e02f97eb87c8a7b9cf407ba1889805a8b471eaf924e70"} Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.220188 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-289lx" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.224305 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a553ef5-b76a-46cd-9e49-30bf4ca86e29-utilities\") pod \"6a553ef5-b76a-46cd-9e49-30bf4ca86e29\" (UID: \"6a553ef5-b76a-46cd-9e49-30bf4ca86e29\") " Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.224700 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a553ef5-b76a-46cd-9e49-30bf4ca86e29-catalog-content\") pod \"6a553ef5-b76a-46cd-9e49-30bf4ca86e29\" (UID: \"6a553ef5-b76a-46cd-9e49-30bf4ca86e29\") " Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.224529 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kp6t2" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.224831 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kv5hk\" (UniqueName: \"kubernetes.io/projected/6a553ef5-b76a-46cd-9e49-30bf4ca86e29-kube-api-access-kv5hk\") pod \"6a553ef5-b76a-46cd-9e49-30bf4ca86e29\" (UID: \"6a553ef5-b76a-46cd-9e49-30bf4ca86e29\") " Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.225297 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76073738-387a-4968-a483-8880e866d1e4-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.225310 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2fac5b3-c60b-4c96-8841-ae53a24aaed1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.225342 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k59xj\" (UniqueName: \"kubernetes.io/projected/e2fac5b3-c60b-4c96-8841-ae53a24aaed1-kube-api-access-k59xj\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.225352 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76073738-387a-4968-a483-8880e866d1e4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.225360 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2fac5b3-c60b-4c96-8841-ae53a24aaed1-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.225369 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqpfc\" (UniqueName: \"kubernetes.io/projected/76073738-387a-4968-a483-8880e866d1e4-kube-api-access-nqpfc\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.226357 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a553ef5-b76a-46cd-9e49-30bf4ca86e29-utilities" (OuterVolumeSpecName: "utilities") pod "6a553ef5-b76a-46cd-9e49-30bf4ca86e29" (UID: "6a553ef5-b76a-46cd-9e49-30bf4ca86e29"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.243412 4813 scope.go:117] "RemoveContainer" containerID="1bb1bb72773c10356ac5e2a9f3183c87652f6b8c98b8b7524252529c7813f497" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.245647 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7d6jw" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.250943 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a553ef5-b76a-46cd-9e49-30bf4ca86e29-kube-api-access-kv5hk" (OuterVolumeSpecName: "kube-api-access-kv5hk") pod "6a553ef5-b76a-46cd-9e49-30bf4ca86e29" (UID: "6a553ef5-b76a-46cd-9e49-30bf4ca86e29"). InnerVolumeSpecName "kube-api-access-kv5hk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.277524 4813 scope.go:117] "RemoveContainer" containerID="3239e6c2ed4ba462e7826d0ff5613cabec75c1a9849a92ccc7f70500ea73c40c" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.298227 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-54xzm"] Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.305163 4813 scope.go:117] "RemoveContainer" containerID="58ab538dbec20a657dfd0e11d156896a425cf1cf03ac50cbbb65c4454607e6f2" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.305275 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-54xzm"] Dec 06 15:50:51 crc kubenswrapper[4813]: E1206 15:50:51.308765 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58ab538dbec20a657dfd0e11d156896a425cf1cf03ac50cbbb65c4454607e6f2\": container with ID starting with 58ab538dbec20a657dfd0e11d156896a425cf1cf03ac50cbbb65c4454607e6f2 not found: ID does not exist" containerID="58ab538dbec20a657dfd0e11d156896a425cf1cf03ac50cbbb65c4454607e6f2" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.309124 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58ab538dbec20a657dfd0e11d156896a425cf1cf03ac50cbbb65c4454607e6f2"} err="failed to get container status \"58ab538dbec20a657dfd0e11d156896a425cf1cf03ac50cbbb65c4454607e6f2\": rpc error: code = NotFound desc = could not find container \"58ab538dbec20a657dfd0e11d156896a425cf1cf03ac50cbbb65c4454607e6f2\": container with ID starting with 58ab538dbec20a657dfd0e11d156896a425cf1cf03ac50cbbb65c4454607e6f2 not found: ID does not exist" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.309226 4813 scope.go:117] "RemoveContainer" containerID="1bb1bb72773c10356ac5e2a9f3183c87652f6b8c98b8b7524252529c7813f497" Dec 06 15:50:51 crc kubenswrapper[4813]: E1206 15:50:51.309837 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bb1bb72773c10356ac5e2a9f3183c87652f6b8c98b8b7524252529c7813f497\": container with ID starting with 1bb1bb72773c10356ac5e2a9f3183c87652f6b8c98b8b7524252529c7813f497 not found: ID does not exist" containerID="1bb1bb72773c10356ac5e2a9f3183c87652f6b8c98b8b7524252529c7813f497" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.309940 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bb1bb72773c10356ac5e2a9f3183c87652f6b8c98b8b7524252529c7813f497"} err="failed to get container status \"1bb1bb72773c10356ac5e2a9f3183c87652f6b8c98b8b7524252529c7813f497\": rpc error: code = NotFound desc = could not find container \"1bb1bb72773c10356ac5e2a9f3183c87652f6b8c98b8b7524252529c7813f497\": container with ID starting with 1bb1bb72773c10356ac5e2a9f3183c87652f6b8c98b8b7524252529c7813f497 not found: ID does not exist" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.310044 4813 scope.go:117] "RemoveContainer" containerID="3239e6c2ed4ba462e7826d0ff5613cabec75c1a9849a92ccc7f70500ea73c40c" Dec 06 15:50:51 crc kubenswrapper[4813]: E1206 15:50:51.310382 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3239e6c2ed4ba462e7826d0ff5613cabec75c1a9849a92ccc7f70500ea73c40c\": container with ID starting with 3239e6c2ed4ba462e7826d0ff5613cabec75c1a9849a92ccc7f70500ea73c40c not found: ID does not exist" containerID="3239e6c2ed4ba462e7826d0ff5613cabec75c1a9849a92ccc7f70500ea73c40c" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.310485 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3239e6c2ed4ba462e7826d0ff5613cabec75c1a9849a92ccc7f70500ea73c40c"} err="failed to get container status \"3239e6c2ed4ba462e7826d0ff5613cabec75c1a9849a92ccc7f70500ea73c40c\": rpc error: code = NotFound desc = could not find container \"3239e6c2ed4ba462e7826d0ff5613cabec75c1a9849a92ccc7f70500ea73c40c\": container with ID starting with 3239e6c2ed4ba462e7826d0ff5613cabec75c1a9849a92ccc7f70500ea73c40c not found: ID does not exist" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.310589 4813 scope.go:117] "RemoveContainer" containerID="6041b776a12e9bbefe8831d00231faeec17ad0e52b7045f06b8132aa5d850c9c" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.326135 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/60b3916a-79fc-42b6-86c6-4cae051457ca-marketplace-trusted-ca\") pod \"60b3916a-79fc-42b6-86c6-4cae051457ca\" (UID: \"60b3916a-79fc-42b6-86c6-4cae051457ca\") " Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.326203 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4z8q\" (UniqueName: \"kubernetes.io/projected/60b3916a-79fc-42b6-86c6-4cae051457ca-kube-api-access-r4z8q\") pod \"60b3916a-79fc-42b6-86c6-4cae051457ca\" (UID: \"60b3916a-79fc-42b6-86c6-4cae051457ca\") " Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.326225 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e982bb16-5f82-418b-8c03-8c73d22e010b-utilities\") pod \"e982bb16-5f82-418b-8c03-8c73d22e010b\" (UID: \"e982bb16-5f82-418b-8c03-8c73d22e010b\") " Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.326289 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpc2w\" (UniqueName: \"kubernetes.io/projected/e982bb16-5f82-418b-8c03-8c73d22e010b-kube-api-access-vpc2w\") pod \"e982bb16-5f82-418b-8c03-8c73d22e010b\" (UID: \"e982bb16-5f82-418b-8c03-8c73d22e010b\") " Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.326322 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da4de6a4-9efc-425c-8712-e66919babdf2-utilities\") pod \"da4de6a4-9efc-425c-8712-e66919babdf2\" (UID: \"da4de6a4-9efc-425c-8712-e66919babdf2\") " Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.326338 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da4de6a4-9efc-425c-8712-e66919babdf2-catalog-content\") pod \"da4de6a4-9efc-425c-8712-e66919babdf2\" (UID: \"da4de6a4-9efc-425c-8712-e66919babdf2\") " Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.326355 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/60b3916a-79fc-42b6-86c6-4cae051457ca-marketplace-operator-metrics\") pod \"60b3916a-79fc-42b6-86c6-4cae051457ca\" (UID: \"60b3916a-79fc-42b6-86c6-4cae051457ca\") " Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.326376 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e982bb16-5f82-418b-8c03-8c73d22e010b-catalog-content\") pod \"e982bb16-5f82-418b-8c03-8c73d22e010b\" (UID: \"e982bb16-5f82-418b-8c03-8c73d22e010b\") " Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.326393 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xft4q\" (UniqueName: \"kubernetes.io/projected/da4de6a4-9efc-425c-8712-e66919babdf2-kube-api-access-xft4q\") pod \"da4de6a4-9efc-425c-8712-e66919babdf2\" (UID: \"da4de6a4-9efc-425c-8712-e66919babdf2\") " Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.326624 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a553ef5-b76a-46cd-9e49-30bf4ca86e29-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.326636 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kv5hk\" (UniqueName: \"kubernetes.io/projected/6a553ef5-b76a-46cd-9e49-30bf4ca86e29-kube-api-access-kv5hk\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.329513 4813 scope.go:117] "RemoveContainer" containerID="27bbb547810dbc893ccfc765f55472dc563d743bb19c743f7e039356935dc817" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.330007 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da4de6a4-9efc-425c-8712-e66919babdf2-kube-api-access-xft4q" (OuterVolumeSpecName: "kube-api-access-xft4q") pod "da4de6a4-9efc-425c-8712-e66919babdf2" (UID: "da4de6a4-9efc-425c-8712-e66919babdf2"). InnerVolumeSpecName "kube-api-access-xft4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.330402 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e982bb16-5f82-418b-8c03-8c73d22e010b-utilities" (OuterVolumeSpecName: "utilities") pod "e982bb16-5f82-418b-8c03-8c73d22e010b" (UID: "e982bb16-5f82-418b-8c03-8c73d22e010b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.331421 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da4de6a4-9efc-425c-8712-e66919babdf2-utilities" (OuterVolumeSpecName: "utilities") pod "da4de6a4-9efc-425c-8712-e66919babdf2" (UID: "da4de6a4-9efc-425c-8712-e66919babdf2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.333917 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60b3916a-79fc-42b6-86c6-4cae051457ca-kube-api-access-r4z8q" (OuterVolumeSpecName: "kube-api-access-r4z8q") pod "60b3916a-79fc-42b6-86c6-4cae051457ca" (UID: "60b3916a-79fc-42b6-86c6-4cae051457ca"). InnerVolumeSpecName "kube-api-access-r4z8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.334318 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60b3916a-79fc-42b6-86c6-4cae051457ca-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "60b3916a-79fc-42b6-86c6-4cae051457ca" (UID: "60b3916a-79fc-42b6-86c6-4cae051457ca"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.335778 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e982bb16-5f82-418b-8c03-8c73d22e010b-kube-api-access-vpc2w" (OuterVolumeSpecName: "kube-api-access-vpc2w") pod "e982bb16-5f82-418b-8c03-8c73d22e010b" (UID: "e982bb16-5f82-418b-8c03-8c73d22e010b"). InnerVolumeSpecName "kube-api-access-vpc2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.335967 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6rwpr"] Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.336626 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60b3916a-79fc-42b6-86c6-4cae051457ca-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "60b3916a-79fc-42b6-86c6-4cae051457ca" (UID: "60b3916a-79fc-42b6-86c6-4cae051457ca"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.339997 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6rwpr"] Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.345950 4813 scope.go:117] "RemoveContainer" containerID="9ddb470f673e893e2ff4c07442e80dc9b2c05b178f3fe952c272878d080a1bdb" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.348848 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da4de6a4-9efc-425c-8712-e66919babdf2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "da4de6a4-9efc-425c-8712-e66919babdf2" (UID: "da4de6a4-9efc-425c-8712-e66919babdf2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.360700 4813 scope.go:117] "RemoveContainer" containerID="5ff03ddb8621464af047b3a4ddbde8a63a7fd6cf544c0e8e25ca07066341a555" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.369445 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a553ef5-b76a-46cd-9e49-30bf4ca86e29-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6a553ef5-b76a-46cd-9e49-30bf4ca86e29" (UID: "6a553ef5-b76a-46cd-9e49-30bf4ca86e29"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.371403 4813 scope.go:117] "RemoveContainer" containerID="de8103356b82043f938ff194c779e0f049cb25beacc418f3491629dc29cadf88" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.384630 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e982bb16-5f82-418b-8c03-8c73d22e010b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e982bb16-5f82-418b-8c03-8c73d22e010b" (UID: "e982bb16-5f82-418b-8c03-8c73d22e010b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.392200 4813 scope.go:117] "RemoveContainer" containerID="88d68d018baeb0bf77b1b600c7fd176ca919cf44a090702ca393ca3c59fec7bd" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.405575 4813 scope.go:117] "RemoveContainer" containerID="5ff03ddb8621464af047b3a4ddbde8a63a7fd6cf544c0e8e25ca07066341a555" Dec 06 15:50:51 crc kubenswrapper[4813]: E1206 15:50:51.406005 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ff03ddb8621464af047b3a4ddbde8a63a7fd6cf544c0e8e25ca07066341a555\": container with ID starting with 5ff03ddb8621464af047b3a4ddbde8a63a7fd6cf544c0e8e25ca07066341a555 not found: ID does not exist" containerID="5ff03ddb8621464af047b3a4ddbde8a63a7fd6cf544c0e8e25ca07066341a555" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.406037 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ff03ddb8621464af047b3a4ddbde8a63a7fd6cf544c0e8e25ca07066341a555"} err="failed to get container status \"5ff03ddb8621464af047b3a4ddbde8a63a7fd6cf544c0e8e25ca07066341a555\": rpc error: code = NotFound desc = could not find container \"5ff03ddb8621464af047b3a4ddbde8a63a7fd6cf544c0e8e25ca07066341a555\": container with ID starting with 5ff03ddb8621464af047b3a4ddbde8a63a7fd6cf544c0e8e25ca07066341a555 not found: ID does not exist" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.406059 4813 scope.go:117] "RemoveContainer" containerID="de8103356b82043f938ff194c779e0f049cb25beacc418f3491629dc29cadf88" Dec 06 15:50:51 crc kubenswrapper[4813]: E1206 15:50:51.406552 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de8103356b82043f938ff194c779e0f049cb25beacc418f3491629dc29cadf88\": container with ID starting with de8103356b82043f938ff194c779e0f049cb25beacc418f3491629dc29cadf88 not found: ID does not exist" containerID="de8103356b82043f938ff194c779e0f049cb25beacc418f3491629dc29cadf88" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.406577 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de8103356b82043f938ff194c779e0f049cb25beacc418f3491629dc29cadf88"} err="failed to get container status \"de8103356b82043f938ff194c779e0f049cb25beacc418f3491629dc29cadf88\": rpc error: code = NotFound desc = could not find container \"de8103356b82043f938ff194c779e0f049cb25beacc418f3491629dc29cadf88\": container with ID starting with de8103356b82043f938ff194c779e0f049cb25beacc418f3491629dc29cadf88 not found: ID does not exist" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.406590 4813 scope.go:117] "RemoveContainer" containerID="88d68d018baeb0bf77b1b600c7fd176ca919cf44a090702ca393ca3c59fec7bd" Dec 06 15:50:51 crc kubenswrapper[4813]: E1206 15:50:51.406856 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88d68d018baeb0bf77b1b600c7fd176ca919cf44a090702ca393ca3c59fec7bd\": container with ID starting with 88d68d018baeb0bf77b1b600c7fd176ca919cf44a090702ca393ca3c59fec7bd not found: ID does not exist" containerID="88d68d018baeb0bf77b1b600c7fd176ca919cf44a090702ca393ca3c59fec7bd" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.406876 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88d68d018baeb0bf77b1b600c7fd176ca919cf44a090702ca393ca3c59fec7bd"} err="failed to get container status \"88d68d018baeb0bf77b1b600c7fd176ca919cf44a090702ca393ca3c59fec7bd\": rpc error: code = NotFound desc = could not find container \"88d68d018baeb0bf77b1b600c7fd176ca919cf44a090702ca393ca3c59fec7bd\": container with ID starting with 88d68d018baeb0bf77b1b600c7fd176ca919cf44a090702ca393ca3c59fec7bd not found: ID does not exist" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.406889 4813 scope.go:117] "RemoveContainer" containerID="dc210c0684fff6a7e1dbf1d2841de30e46e582fb956d958e6cb6ed48c0837da8" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.418299 4813 scope.go:117] "RemoveContainer" containerID="cd856a72ba6a857e1551e16c291ffcd37c2f23cf257cb4aceaa7b977c234eef8" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.427423 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/952e1f68-97e3-4191-9dde-43d67e83a1e0-catalog-content\") pod \"952e1f68-97e3-4191-9dde-43d67e83a1e0\" (UID: \"952e1f68-97e3-4191-9dde-43d67e83a1e0\") " Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.427485 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/952e1f68-97e3-4191-9dde-43d67e83a1e0-utilities\") pod \"952e1f68-97e3-4191-9dde-43d67e83a1e0\" (UID: \"952e1f68-97e3-4191-9dde-43d67e83a1e0\") " Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.427547 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqpc5\" (UniqueName: \"kubernetes.io/projected/952e1f68-97e3-4191-9dde-43d67e83a1e0-kube-api-access-pqpc5\") pod \"952e1f68-97e3-4191-9dde-43d67e83a1e0\" (UID: \"952e1f68-97e3-4191-9dde-43d67e83a1e0\") " Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.427737 4813 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/60b3916a-79fc-42b6-86c6-4cae051457ca-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.427753 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4z8q\" (UniqueName: \"kubernetes.io/projected/60b3916a-79fc-42b6-86c6-4cae051457ca-kube-api-access-r4z8q\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.427763 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e982bb16-5f82-418b-8c03-8c73d22e010b-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.427775 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpc2w\" (UniqueName: \"kubernetes.io/projected/e982bb16-5f82-418b-8c03-8c73d22e010b-kube-api-access-vpc2w\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.427786 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a553ef5-b76a-46cd-9e49-30bf4ca86e29-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.427796 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da4de6a4-9efc-425c-8712-e66919babdf2-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.427803 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da4de6a4-9efc-425c-8712-e66919babdf2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.427815 4813 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/60b3916a-79fc-42b6-86c6-4cae051457ca-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.427827 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e982bb16-5f82-418b-8c03-8c73d22e010b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.427838 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xft4q\" (UniqueName: \"kubernetes.io/projected/da4de6a4-9efc-425c-8712-e66919babdf2-kube-api-access-xft4q\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.428623 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/952e1f68-97e3-4191-9dde-43d67e83a1e0-utilities" (OuterVolumeSpecName: "utilities") pod "952e1f68-97e3-4191-9dde-43d67e83a1e0" (UID: "952e1f68-97e3-4191-9dde-43d67e83a1e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.430880 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/952e1f68-97e3-4191-9dde-43d67e83a1e0-kube-api-access-pqpc5" (OuterVolumeSpecName: "kube-api-access-pqpc5") pod "952e1f68-97e3-4191-9dde-43d67e83a1e0" (UID: "952e1f68-97e3-4191-9dde-43d67e83a1e0"). InnerVolumeSpecName "kube-api-access-pqpc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.434540 4813 scope.go:117] "RemoveContainer" containerID="bc96f70a14b2fbc8f38437bcc68ace07662e44f0bf69a3bb3a0824987b67e5a1" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.450470 4813 scope.go:117] "RemoveContainer" containerID="dc210c0684fff6a7e1dbf1d2841de30e46e582fb956d958e6cb6ed48c0837da8" Dec 06 15:50:51 crc kubenswrapper[4813]: E1206 15:50:51.450952 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc210c0684fff6a7e1dbf1d2841de30e46e582fb956d958e6cb6ed48c0837da8\": container with ID starting with dc210c0684fff6a7e1dbf1d2841de30e46e582fb956d958e6cb6ed48c0837da8 not found: ID does not exist" containerID="dc210c0684fff6a7e1dbf1d2841de30e46e582fb956d958e6cb6ed48c0837da8" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.451003 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc210c0684fff6a7e1dbf1d2841de30e46e582fb956d958e6cb6ed48c0837da8"} err="failed to get container status \"dc210c0684fff6a7e1dbf1d2841de30e46e582fb956d958e6cb6ed48c0837da8\": rpc error: code = NotFound desc = could not find container \"dc210c0684fff6a7e1dbf1d2841de30e46e582fb956d958e6cb6ed48c0837da8\": container with ID starting with dc210c0684fff6a7e1dbf1d2841de30e46e582fb956d958e6cb6ed48c0837da8 not found: ID does not exist" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.451038 4813 scope.go:117] "RemoveContainer" containerID="cd856a72ba6a857e1551e16c291ffcd37c2f23cf257cb4aceaa7b977c234eef8" Dec 06 15:50:51 crc kubenswrapper[4813]: E1206 15:50:51.451530 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd856a72ba6a857e1551e16c291ffcd37c2f23cf257cb4aceaa7b977c234eef8\": container with ID starting with cd856a72ba6a857e1551e16c291ffcd37c2f23cf257cb4aceaa7b977c234eef8 not found: ID does not exist" containerID="cd856a72ba6a857e1551e16c291ffcd37c2f23cf257cb4aceaa7b977c234eef8" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.451572 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd856a72ba6a857e1551e16c291ffcd37c2f23cf257cb4aceaa7b977c234eef8"} err="failed to get container status \"cd856a72ba6a857e1551e16c291ffcd37c2f23cf257cb4aceaa7b977c234eef8\": rpc error: code = NotFound desc = could not find container \"cd856a72ba6a857e1551e16c291ffcd37c2f23cf257cb4aceaa7b977c234eef8\": container with ID starting with cd856a72ba6a857e1551e16c291ffcd37c2f23cf257cb4aceaa7b977c234eef8 not found: ID does not exist" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.451597 4813 scope.go:117] "RemoveContainer" containerID="bc96f70a14b2fbc8f38437bcc68ace07662e44f0bf69a3bb3a0824987b67e5a1" Dec 06 15:50:51 crc kubenswrapper[4813]: E1206 15:50:51.451969 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc96f70a14b2fbc8f38437bcc68ace07662e44f0bf69a3bb3a0824987b67e5a1\": container with ID starting with bc96f70a14b2fbc8f38437bcc68ace07662e44f0bf69a3bb3a0824987b67e5a1 not found: ID does not exist" containerID="bc96f70a14b2fbc8f38437bcc68ace07662e44f0bf69a3bb3a0824987b67e5a1" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.452020 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc96f70a14b2fbc8f38437bcc68ace07662e44f0bf69a3bb3a0824987b67e5a1"} err="failed to get container status \"bc96f70a14b2fbc8f38437bcc68ace07662e44f0bf69a3bb3a0824987b67e5a1\": rpc error: code = NotFound desc = could not find container \"bc96f70a14b2fbc8f38437bcc68ace07662e44f0bf69a3bb3a0824987b67e5a1\": container with ID starting with bc96f70a14b2fbc8f38437bcc68ace07662e44f0bf69a3bb3a0824987b67e5a1 not found: ID does not exist" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.537769 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/952e1f68-97e3-4191-9dde-43d67e83a1e0-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.537811 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqpc5\" (UniqueName: \"kubernetes.io/projected/952e1f68-97e3-4191-9dde-43d67e83a1e0-kube-api-access-pqpc5\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.549778 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/952e1f68-97e3-4191-9dde-43d67e83a1e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "952e1f68-97e3-4191-9dde-43d67e83a1e0" (UID: "952e1f68-97e3-4191-9dde-43d67e83a1e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.556911 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vn7fl"] Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.562092 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vn7fl"] Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.588909 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5zp6r"] Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.592233 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5zp6r"] Dec 06 15:50:51 crc kubenswrapper[4813]: I1206 15:50:51.638875 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/952e1f68-97e3-4191-9dde-43d67e83a1e0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.128101 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.206716 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.206783 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.227514 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7d6jw" event={"ID":"952e1f68-97e3-4191-9dde-43d67e83a1e0","Type":"ContainerDied","Data":"a22f4bb1238a969321aed571002743042e17a097213d53ae9ab79abd34b9da6c"} Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.227551 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7d6jw" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.227577 4813 scope.go:117] "RemoveContainer" containerID="fc5c2d08cc45d223b3dddb5282d706c77ec8854aa99c6a85001fec025351536f" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.230983 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kp6t2" event={"ID":"e982bb16-5f82-418b-8c03-8c73d22e010b","Type":"ContainerDied","Data":"dc06103c0a720defb5a411399c5900332373d97efcbe58424f0879cd6c72645b"} Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.231084 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kp6t2" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.232513 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-289lx" event={"ID":"60b3916a-79fc-42b6-86c6-4cae051457ca","Type":"ContainerDied","Data":"0437e6a102086e57333e6f9f9d3885373f64e275fc8466ca2c7ad01c94bd06f1"} Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.232545 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-289lx" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.253209 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.253247 4813 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="0eb54412ce9438f87b8eb1721001f4d5bb20f02f364dc78f126722417786f8ac" exitCode=137 Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.253355 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.267527 4813 scope.go:117] "RemoveContainer" containerID="5c912bf5b91f301bd9b0351d051dfb46642aa95008de5ccb5206464e915fc5e5" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.274559 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7d6jw"] Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.277701 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7d6jw"] Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.289070 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-289lx"] Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.291642 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-289lx"] Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.309405 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kp6t2"] Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.313346 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kp6t2"] Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.313599 4813 scope.go:117] "RemoveContainer" containerID="f92e84ded684cbb4d8495f28050e654be29b0c5ea9192efc0803cba604d48054" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.332610 4813 scope.go:117] "RemoveContainer" containerID="6e4cbadc5a42dd6f512afa67014f67d6c1c96d3dbe1b330a18894f6c5d98b8d3" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.350297 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.350362 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.350383 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.350402 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.350422 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.350480 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.350556 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.350627 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.350697 4813 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.350718 4813 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.350763 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.351692 4813 scope.go:117] "RemoveContainer" containerID="009b5e53cb626741235644fa6abe27c1d2b476b93cdd1de42ae8b6e277c6d5bb" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.360302 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.366645 4813 scope.go:117] "RemoveContainer" containerID="a76753c7cb31be3afa135159cb50365b7cb5e7f2c369b2f81b2b7e4a7b2deb76" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.382491 4813 scope.go:117] "RemoveContainer" containerID="441756359cdf72b3e97cab1b8ce564665e57f14ec2279db9364d1706faa70be5" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.400292 4813 scope.go:117] "RemoveContainer" containerID="0eb54412ce9438f87b8eb1721001f4d5bb20f02f364dc78f126722417786f8ac" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.411339 4813 scope.go:117] "RemoveContainer" containerID="0eb54412ce9438f87b8eb1721001f4d5bb20f02f364dc78f126722417786f8ac" Dec 06 15:50:52 crc kubenswrapper[4813]: E1206 15:50:52.411670 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0eb54412ce9438f87b8eb1721001f4d5bb20f02f364dc78f126722417786f8ac\": container with ID starting with 0eb54412ce9438f87b8eb1721001f4d5bb20f02f364dc78f126722417786f8ac not found: ID does not exist" containerID="0eb54412ce9438f87b8eb1721001f4d5bb20f02f364dc78f126722417786f8ac" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.411701 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0eb54412ce9438f87b8eb1721001f4d5bb20f02f364dc78f126722417786f8ac"} err="failed to get container status \"0eb54412ce9438f87b8eb1721001f4d5bb20f02f364dc78f126722417786f8ac\": rpc error: code = NotFound desc = could not find container \"0eb54412ce9438f87b8eb1721001f4d5bb20f02f364dc78f126722417786f8ac\": container with ID starting with 0eb54412ce9438f87b8eb1721001f4d5bb20f02f364dc78f126722417786f8ac not found: ID does not exist" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.451766 4813 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.451818 4813 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.451854 4813 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.497117 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60b3916a-79fc-42b6-86c6-4cae051457ca" path="/var/lib/kubelet/pods/60b3916a-79fc-42b6-86c6-4cae051457ca/volumes" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.498467 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a553ef5-b76a-46cd-9e49-30bf4ca86e29" path="/var/lib/kubelet/pods/6a553ef5-b76a-46cd-9e49-30bf4ca86e29/volumes" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.500088 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76073738-387a-4968-a483-8880e866d1e4" path="/var/lib/kubelet/pods/76073738-387a-4968-a483-8880e866d1e4/volumes" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.502372 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="952e1f68-97e3-4191-9dde-43d67e83a1e0" path="/var/lib/kubelet/pods/952e1f68-97e3-4191-9dde-43d67e83a1e0/volumes" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.503741 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da4de6a4-9efc-425c-8712-e66919babdf2" path="/var/lib/kubelet/pods/da4de6a4-9efc-425c-8712-e66919babdf2/volumes" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.506341 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2fac5b3-c60b-4c96-8841-ae53a24aaed1" path="/var/lib/kubelet/pods/e2fac5b3-c60b-4c96-8841-ae53a24aaed1/volumes" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.507671 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e982bb16-5f82-418b-8c03-8c73d22e010b" path="/var/lib/kubelet/pods/e982bb16-5f82-418b-8c03-8c73d22e010b/volumes" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.508632 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.524637 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 06 15:50:52 crc kubenswrapper[4813]: I1206 15:50:52.824485 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 06 15:50:53 crc kubenswrapper[4813]: I1206 15:50:53.101165 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 06 15:50:54 crc kubenswrapper[4813]: I1206 15:50:54.105699 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 06 15:51:00 crc kubenswrapper[4813]: I1206 15:51:00.329988 4813 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.299951 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-v6fmp"] Dec 06 15:51:04 crc kubenswrapper[4813]: E1206 15:51:04.300522 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2fac5b3-c60b-4c96-8841-ae53a24aaed1" containerName="extract-content" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300533 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2fac5b3-c60b-4c96-8841-ae53a24aaed1" containerName="extract-content" Dec 06 15:51:04 crc kubenswrapper[4813]: E1206 15:51:04.300543 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="952e1f68-97e3-4191-9dde-43d67e83a1e0" containerName="extract-utilities" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300549 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="952e1f68-97e3-4191-9dde-43d67e83a1e0" containerName="extract-utilities" Dec 06 15:51:04 crc kubenswrapper[4813]: E1206 15:51:04.300557 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da4de6a4-9efc-425c-8712-e66919babdf2" containerName="extract-content" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300563 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="da4de6a4-9efc-425c-8712-e66919babdf2" containerName="extract-content" Dec 06 15:51:04 crc kubenswrapper[4813]: E1206 15:51:04.300572 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60b3916a-79fc-42b6-86c6-4cae051457ca" containerName="marketplace-operator" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300577 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="60b3916a-79fc-42b6-86c6-4cae051457ca" containerName="marketplace-operator" Dec 06 15:51:04 crc kubenswrapper[4813]: E1206 15:51:04.300584 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e982bb16-5f82-418b-8c03-8c73d22e010b" containerName="registry-server" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300590 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e982bb16-5f82-418b-8c03-8c73d22e010b" containerName="registry-server" Dec 06 15:51:04 crc kubenswrapper[4813]: E1206 15:51:04.300598 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2fac5b3-c60b-4c96-8841-ae53a24aaed1" containerName="extract-utilities" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300603 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2fac5b3-c60b-4c96-8841-ae53a24aaed1" containerName="extract-utilities" Dec 06 15:51:04 crc kubenswrapper[4813]: E1206 15:51:04.300612 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a553ef5-b76a-46cd-9e49-30bf4ca86e29" containerName="extract-utilities" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300618 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a553ef5-b76a-46cd-9e49-30bf4ca86e29" containerName="extract-utilities" Dec 06 15:51:04 crc kubenswrapper[4813]: E1206 15:51:04.300625 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76073738-387a-4968-a483-8880e866d1e4" containerName="extract-utilities" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300630 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="76073738-387a-4968-a483-8880e866d1e4" containerName="extract-utilities" Dec 06 15:51:04 crc kubenswrapper[4813]: E1206 15:51:04.300640 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2fac5b3-c60b-4c96-8841-ae53a24aaed1" containerName="registry-server" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300645 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2fac5b3-c60b-4c96-8841-ae53a24aaed1" containerName="registry-server" Dec 06 15:51:04 crc kubenswrapper[4813]: E1206 15:51:04.300653 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a553ef5-b76a-46cd-9e49-30bf4ca86e29" containerName="registry-server" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300660 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a553ef5-b76a-46cd-9e49-30bf4ca86e29" containerName="registry-server" Dec 06 15:51:04 crc kubenswrapper[4813]: E1206 15:51:04.300667 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e982bb16-5f82-418b-8c03-8c73d22e010b" containerName="extract-content" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300674 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e982bb16-5f82-418b-8c03-8c73d22e010b" containerName="extract-content" Dec 06 15:51:04 crc kubenswrapper[4813]: E1206 15:51:04.300680 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da4de6a4-9efc-425c-8712-e66919babdf2" containerName="extract-utilities" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300686 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="da4de6a4-9efc-425c-8712-e66919babdf2" containerName="extract-utilities" Dec 06 15:51:04 crc kubenswrapper[4813]: E1206 15:51:04.300694 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="952e1f68-97e3-4191-9dde-43d67e83a1e0" containerName="extract-content" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300700 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="952e1f68-97e3-4191-9dde-43d67e83a1e0" containerName="extract-content" Dec 06 15:51:04 crc kubenswrapper[4813]: E1206 15:51:04.300706 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f522807f-660d-4791-95f4-f2d9daa58936" containerName="extract-utilities" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300712 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f522807f-660d-4791-95f4-f2d9daa58936" containerName="extract-utilities" Dec 06 15:51:04 crc kubenswrapper[4813]: E1206 15:51:04.300718 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76073738-387a-4968-a483-8880e866d1e4" containerName="registry-server" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300724 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="76073738-387a-4968-a483-8880e866d1e4" containerName="registry-server" Dec 06 15:51:04 crc kubenswrapper[4813]: E1206 15:51:04.300732 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a553ef5-b76a-46cd-9e49-30bf4ca86e29" containerName="extract-content" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300737 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a553ef5-b76a-46cd-9e49-30bf4ca86e29" containerName="extract-content" Dec 06 15:51:04 crc kubenswrapper[4813]: E1206 15:51:04.300744 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da4de6a4-9efc-425c-8712-e66919babdf2" containerName="registry-server" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300750 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="da4de6a4-9efc-425c-8712-e66919babdf2" containerName="registry-server" Dec 06 15:51:04 crc kubenswrapper[4813]: E1206 15:51:04.300759 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f522807f-660d-4791-95f4-f2d9daa58936" containerName="registry-server" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300765 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f522807f-660d-4791-95f4-f2d9daa58936" containerName="registry-server" Dec 06 15:51:04 crc kubenswrapper[4813]: E1206 15:51:04.300773 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff" containerName="installer" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300778 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff" containerName="installer" Dec 06 15:51:04 crc kubenswrapper[4813]: E1206 15:51:04.300787 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f522807f-660d-4791-95f4-f2d9daa58936" containerName="extract-content" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300793 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f522807f-660d-4791-95f4-f2d9daa58936" containerName="extract-content" Dec 06 15:51:04 crc kubenswrapper[4813]: E1206 15:51:04.300801 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="952e1f68-97e3-4191-9dde-43d67e83a1e0" containerName="registry-server" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300806 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="952e1f68-97e3-4191-9dde-43d67e83a1e0" containerName="registry-server" Dec 06 15:51:04 crc kubenswrapper[4813]: E1206 15:51:04.300814 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e982bb16-5f82-418b-8c03-8c73d22e010b" containerName="extract-utilities" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300820 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e982bb16-5f82-418b-8c03-8c73d22e010b" containerName="extract-utilities" Dec 06 15:51:04 crc kubenswrapper[4813]: E1206 15:51:04.300828 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76073738-387a-4968-a483-8880e866d1e4" containerName="extract-content" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300834 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="76073738-387a-4968-a483-8880e866d1e4" containerName="extract-content" Dec 06 15:51:04 crc kubenswrapper[4813]: E1206 15:51:04.300841 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300846 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300942 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="76073738-387a-4968-a483-8880e866d1e4" containerName="registry-server" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300950 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="952e1f68-97e3-4191-9dde-43d67e83a1e0" containerName="registry-server" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300960 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ea9fcd6-1a07-4903-a615-3a2e2d01c6ff" containerName="installer" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300967 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a553ef5-b76a-46cd-9e49-30bf4ca86e29" containerName="registry-server" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300977 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300984 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f522807f-660d-4791-95f4-f2d9daa58936" containerName="registry-server" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300990 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="e982bb16-5f82-418b-8c03-8c73d22e010b" containerName="registry-server" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.300996 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="da4de6a4-9efc-425c-8712-e66919babdf2" containerName="registry-server" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.301003 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2fac5b3-c60b-4c96-8841-ae53a24aaed1" containerName="registry-server" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.301010 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="60b3916a-79fc-42b6-86c6-4cae051457ca" containerName="marketplace-operator" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.301324 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-v6fmp" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.308294 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.308491 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.308589 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.310291 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.318177 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-v6fmp"] Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.322421 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.381237 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7lpk\" (UniqueName: \"kubernetes.io/projected/2660f707-b9cc-4c4f-b42a-9aa6699ef326-kube-api-access-c7lpk\") pod \"marketplace-operator-79b997595-v6fmp\" (UID: \"2660f707-b9cc-4c4f-b42a-9aa6699ef326\") " pod="openshift-marketplace/marketplace-operator-79b997595-v6fmp" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.381306 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2660f707-b9cc-4c4f-b42a-9aa6699ef326-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-v6fmp\" (UID: \"2660f707-b9cc-4c4f-b42a-9aa6699ef326\") " pod="openshift-marketplace/marketplace-operator-79b997595-v6fmp" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.381335 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2660f707-b9cc-4c4f-b42a-9aa6699ef326-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-v6fmp\" (UID: \"2660f707-b9cc-4c4f-b42a-9aa6699ef326\") " pod="openshift-marketplace/marketplace-operator-79b997595-v6fmp" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.482638 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7lpk\" (UniqueName: \"kubernetes.io/projected/2660f707-b9cc-4c4f-b42a-9aa6699ef326-kube-api-access-c7lpk\") pod \"marketplace-operator-79b997595-v6fmp\" (UID: \"2660f707-b9cc-4c4f-b42a-9aa6699ef326\") " pod="openshift-marketplace/marketplace-operator-79b997595-v6fmp" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.482684 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2660f707-b9cc-4c4f-b42a-9aa6699ef326-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-v6fmp\" (UID: \"2660f707-b9cc-4c4f-b42a-9aa6699ef326\") " pod="openshift-marketplace/marketplace-operator-79b997595-v6fmp" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.482707 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2660f707-b9cc-4c4f-b42a-9aa6699ef326-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-v6fmp\" (UID: \"2660f707-b9cc-4c4f-b42a-9aa6699ef326\") " pod="openshift-marketplace/marketplace-operator-79b997595-v6fmp" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.483622 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2660f707-b9cc-4c4f-b42a-9aa6699ef326-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-v6fmp\" (UID: \"2660f707-b9cc-4c4f-b42a-9aa6699ef326\") " pod="openshift-marketplace/marketplace-operator-79b997595-v6fmp" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.488038 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2660f707-b9cc-4c4f-b42a-9aa6699ef326-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-v6fmp\" (UID: \"2660f707-b9cc-4c4f-b42a-9aa6699ef326\") " pod="openshift-marketplace/marketplace-operator-79b997595-v6fmp" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.500946 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7lpk\" (UniqueName: \"kubernetes.io/projected/2660f707-b9cc-4c4f-b42a-9aa6699ef326-kube-api-access-c7lpk\") pod \"marketplace-operator-79b997595-v6fmp\" (UID: \"2660f707-b9cc-4c4f-b42a-9aa6699ef326\") " pod="openshift-marketplace/marketplace-operator-79b997595-v6fmp" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.618850 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-v6fmp" Dec 06 15:51:04 crc kubenswrapper[4813]: I1206 15:51:04.819835 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-v6fmp"] Dec 06 15:51:05 crc kubenswrapper[4813]: I1206 15:51:05.747002 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-v6fmp" event={"ID":"2660f707-b9cc-4c4f-b42a-9aa6699ef326","Type":"ContainerStarted","Data":"4c021e708fb5df10ed43d78cd81fa92cb168e4934f03df9d1feeb95aab3992f4"} Dec 06 15:51:05 crc kubenswrapper[4813]: I1206 15:51:05.747581 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-v6fmp" event={"ID":"2660f707-b9cc-4c4f-b42a-9aa6699ef326","Type":"ContainerStarted","Data":"30974318019d4097087acfe239e5cd2df10e207b228fed38ae7bb05ec9fed31f"} Dec 06 15:51:05 crc kubenswrapper[4813]: I1206 15:51:05.747659 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-v6fmp" Dec 06 15:51:05 crc kubenswrapper[4813]: I1206 15:51:05.756292 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-v6fmp" Dec 06 15:51:05 crc kubenswrapper[4813]: I1206 15:51:05.791612 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-v6fmp" podStartSLOduration=1.7915920029999999 podStartE2EDuration="1.791592003s" podCreationTimestamp="2025-12-06 15:51:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:51:05.767107057 +0000 UTC m=+305.657986663" watchObservedRunningTime="2025-12-06 15:51:05.791592003 +0000 UTC m=+305.682471579" Dec 06 15:51:11 crc kubenswrapper[4813]: I1206 15:51:11.774737 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 06 15:51:11 crc kubenswrapper[4813]: I1206 15:51:11.776150 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 06 15:51:11 crc kubenswrapper[4813]: I1206 15:51:11.776175 4813 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="17af73d1dbe8321c83f2bb87f0aff0cb05824a6c180b77b4869bb4fa29cc81b3" exitCode=137 Dec 06 15:51:11 crc kubenswrapper[4813]: I1206 15:51:11.776201 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"17af73d1dbe8321c83f2bb87f0aff0cb05824a6c180b77b4869bb4fa29cc81b3"} Dec 06 15:51:11 crc kubenswrapper[4813]: I1206 15:51:11.776231 4813 scope.go:117] "RemoveContainer" containerID="cf357600f187a6249f7fb14d87b985dd10e42aeca3a49e71bcc201ec5ff0c545" Dec 06 15:51:12 crc kubenswrapper[4813]: I1206 15:51:12.782092 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 06 15:51:12 crc kubenswrapper[4813]: I1206 15:51:12.783197 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"505f20b678278dd1d114edf89df1b5c8158113f2ca1504d5f64bc8ca8401157a"} Dec 06 15:51:21 crc kubenswrapper[4813]: I1206 15:51:21.597335 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 15:51:21 crc kubenswrapper[4813]: I1206 15:51:21.618690 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 15:51:21 crc kubenswrapper[4813]: I1206 15:51:21.828685 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 15:51:21 crc kubenswrapper[4813]: I1206 15:51:21.970949 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xwq7l"] Dec 06 15:51:21 crc kubenswrapper[4813]: I1206 15:51:21.973193 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xwq7l" Dec 06 15:51:21 crc kubenswrapper[4813]: I1206 15:51:21.978870 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 06 15:51:22 crc kubenswrapper[4813]: I1206 15:51:22.000788 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xwq7l"] Dec 06 15:51:22 crc kubenswrapper[4813]: I1206 15:51:22.093031 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca46ce83-a423-478a-9d44-d45a0ef28ef3-catalog-content\") pod \"redhat-marketplace-xwq7l\" (UID: \"ca46ce83-a423-478a-9d44-d45a0ef28ef3\") " pod="openshift-marketplace/redhat-marketplace-xwq7l" Dec 06 15:51:22 crc kubenswrapper[4813]: I1206 15:51:22.093414 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2gs6\" (UniqueName: \"kubernetes.io/projected/ca46ce83-a423-478a-9d44-d45a0ef28ef3-kube-api-access-j2gs6\") pod \"redhat-marketplace-xwq7l\" (UID: \"ca46ce83-a423-478a-9d44-d45a0ef28ef3\") " pod="openshift-marketplace/redhat-marketplace-xwq7l" Dec 06 15:51:22 crc kubenswrapper[4813]: I1206 15:51:22.093536 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca46ce83-a423-478a-9d44-d45a0ef28ef3-utilities\") pod \"redhat-marketplace-xwq7l\" (UID: \"ca46ce83-a423-478a-9d44-d45a0ef28ef3\") " pod="openshift-marketplace/redhat-marketplace-xwq7l" Dec 06 15:51:22 crc kubenswrapper[4813]: I1206 15:51:22.157642 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ws47p"] Dec 06 15:51:22 crc kubenswrapper[4813]: I1206 15:51:22.158640 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ws47p" Dec 06 15:51:22 crc kubenswrapper[4813]: W1206 15:51:22.160949 4813 reflector.go:561] object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g": failed to list *v1.Secret: secrets "certified-operators-dockercfg-4rs5g" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Dec 06 15:51:22 crc kubenswrapper[4813]: E1206 15:51:22.160991 4813 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"certified-operators-dockercfg-4rs5g\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"certified-operators-dockercfg-4rs5g\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 06 15:51:22 crc kubenswrapper[4813]: I1206 15:51:22.183598 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ws47p"] Dec 06 15:51:22 crc kubenswrapper[4813]: I1206 15:51:22.195042 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca46ce83-a423-478a-9d44-d45a0ef28ef3-catalog-content\") pod \"redhat-marketplace-xwq7l\" (UID: \"ca46ce83-a423-478a-9d44-d45a0ef28ef3\") " pod="openshift-marketplace/redhat-marketplace-xwq7l" Dec 06 15:51:22 crc kubenswrapper[4813]: I1206 15:51:22.195097 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2gs6\" (UniqueName: \"kubernetes.io/projected/ca46ce83-a423-478a-9d44-d45a0ef28ef3-kube-api-access-j2gs6\") pod \"redhat-marketplace-xwq7l\" (UID: \"ca46ce83-a423-478a-9d44-d45a0ef28ef3\") " pod="openshift-marketplace/redhat-marketplace-xwq7l" Dec 06 15:51:22 crc kubenswrapper[4813]: I1206 15:51:22.195123 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca46ce83-a423-478a-9d44-d45a0ef28ef3-utilities\") pod \"redhat-marketplace-xwq7l\" (UID: \"ca46ce83-a423-478a-9d44-d45a0ef28ef3\") " pod="openshift-marketplace/redhat-marketplace-xwq7l" Dec 06 15:51:22 crc kubenswrapper[4813]: I1206 15:51:22.195672 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca46ce83-a423-478a-9d44-d45a0ef28ef3-utilities\") pod \"redhat-marketplace-xwq7l\" (UID: \"ca46ce83-a423-478a-9d44-d45a0ef28ef3\") " pod="openshift-marketplace/redhat-marketplace-xwq7l" Dec 06 15:51:22 crc kubenswrapper[4813]: I1206 15:51:22.195684 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca46ce83-a423-478a-9d44-d45a0ef28ef3-catalog-content\") pod \"redhat-marketplace-xwq7l\" (UID: \"ca46ce83-a423-478a-9d44-d45a0ef28ef3\") " pod="openshift-marketplace/redhat-marketplace-xwq7l" Dec 06 15:51:22 crc kubenswrapper[4813]: I1206 15:51:22.222899 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2gs6\" (UniqueName: \"kubernetes.io/projected/ca46ce83-a423-478a-9d44-d45a0ef28ef3-kube-api-access-j2gs6\") pod \"redhat-marketplace-xwq7l\" (UID: \"ca46ce83-a423-478a-9d44-d45a0ef28ef3\") " pod="openshift-marketplace/redhat-marketplace-xwq7l" Dec 06 15:51:22 crc kubenswrapper[4813]: I1206 15:51:22.276658 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 15:51:22 crc kubenswrapper[4813]: I1206 15:51:22.296331 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tv6xm\" (UniqueName: \"kubernetes.io/projected/abeb6c67-8702-4211-9aab-7cdf25c0d078-kube-api-access-tv6xm\") pod \"certified-operators-ws47p\" (UID: \"abeb6c67-8702-4211-9aab-7cdf25c0d078\") " pod="openshift-marketplace/certified-operators-ws47p" Dec 06 15:51:22 crc kubenswrapper[4813]: I1206 15:51:22.296594 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abeb6c67-8702-4211-9aab-7cdf25c0d078-catalog-content\") pod \"certified-operators-ws47p\" (UID: \"abeb6c67-8702-4211-9aab-7cdf25c0d078\") " pod="openshift-marketplace/certified-operators-ws47p" Dec 06 15:51:22 crc kubenswrapper[4813]: I1206 15:51:22.296725 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abeb6c67-8702-4211-9aab-7cdf25c0d078-utilities\") pod \"certified-operators-ws47p\" (UID: \"abeb6c67-8702-4211-9aab-7cdf25c0d078\") " pod="openshift-marketplace/certified-operators-ws47p" Dec 06 15:51:22 crc kubenswrapper[4813]: I1206 15:51:22.305316 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xwq7l" Dec 06 15:51:22 crc kubenswrapper[4813]: I1206 15:51:22.398612 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abeb6c67-8702-4211-9aab-7cdf25c0d078-catalog-content\") pod \"certified-operators-ws47p\" (UID: \"abeb6c67-8702-4211-9aab-7cdf25c0d078\") " pod="openshift-marketplace/certified-operators-ws47p" Dec 06 15:51:22 crc kubenswrapper[4813]: I1206 15:51:22.399044 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abeb6c67-8702-4211-9aab-7cdf25c0d078-utilities\") pod \"certified-operators-ws47p\" (UID: \"abeb6c67-8702-4211-9aab-7cdf25c0d078\") " pod="openshift-marketplace/certified-operators-ws47p" Dec 06 15:51:22 crc kubenswrapper[4813]: I1206 15:51:22.399137 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tv6xm\" (UniqueName: \"kubernetes.io/projected/abeb6c67-8702-4211-9aab-7cdf25c0d078-kube-api-access-tv6xm\") pod \"certified-operators-ws47p\" (UID: \"abeb6c67-8702-4211-9aab-7cdf25c0d078\") " pod="openshift-marketplace/certified-operators-ws47p" Dec 06 15:51:22 crc kubenswrapper[4813]: I1206 15:51:22.399737 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abeb6c67-8702-4211-9aab-7cdf25c0d078-catalog-content\") pod \"certified-operators-ws47p\" (UID: \"abeb6c67-8702-4211-9aab-7cdf25c0d078\") " pod="openshift-marketplace/certified-operators-ws47p" Dec 06 15:51:22 crc kubenswrapper[4813]: I1206 15:51:22.399829 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abeb6c67-8702-4211-9aab-7cdf25c0d078-utilities\") pod \"certified-operators-ws47p\" (UID: \"abeb6c67-8702-4211-9aab-7cdf25c0d078\") " pod="openshift-marketplace/certified-operators-ws47p" Dec 06 15:51:22 crc kubenswrapper[4813]: I1206 15:51:22.449251 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tv6xm\" (UniqueName: \"kubernetes.io/projected/abeb6c67-8702-4211-9aab-7cdf25c0d078-kube-api-access-tv6xm\") pod \"certified-operators-ws47p\" (UID: \"abeb6c67-8702-4211-9aab-7cdf25c0d078\") " pod="openshift-marketplace/certified-operators-ws47p" Dec 06 15:51:22 crc kubenswrapper[4813]: I1206 15:51:22.761598 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xwq7l"] Dec 06 15:51:22 crc kubenswrapper[4813]: W1206 15:51:22.780445 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca46ce83_a423_478a_9d44_d45a0ef28ef3.slice/crio-c9fd1ee34ce2902197ada4aab54153c0b685433d6db8d8d5dd2b3a7156459f3c WatchSource:0}: Error finding container c9fd1ee34ce2902197ada4aab54153c0b685433d6db8d8d5dd2b3a7156459f3c: Status 404 returned error can't find the container with id c9fd1ee34ce2902197ada4aab54153c0b685433d6db8d8d5dd2b3a7156459f3c Dec 06 15:51:22 crc kubenswrapper[4813]: I1206 15:51:22.839907 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xwq7l" event={"ID":"ca46ce83-a423-478a-9d44-d45a0ef28ef3","Type":"ContainerStarted","Data":"c9fd1ee34ce2902197ada4aab54153c0b685433d6db8d8d5dd2b3a7156459f3c"} Dec 06 15:51:23 crc kubenswrapper[4813]: I1206 15:51:23.222240 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 06 15:51:23 crc kubenswrapper[4813]: I1206 15:51:23.231774 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ws47p" Dec 06 15:51:23 crc kubenswrapper[4813]: I1206 15:51:23.550710 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ws47p"] Dec 06 15:51:23 crc kubenswrapper[4813]: I1206 15:51:23.848593 4813 generic.go:334] "Generic (PLEG): container finished" podID="ca46ce83-a423-478a-9d44-d45a0ef28ef3" containerID="3c3ae12b5608e80f0cc359c9a34178c377bf55e20c980e0ad38d2e6a086de25c" exitCode=0 Dec 06 15:51:23 crc kubenswrapper[4813]: I1206 15:51:23.849765 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xwq7l" event={"ID":"ca46ce83-a423-478a-9d44-d45a0ef28ef3","Type":"ContainerDied","Data":"3c3ae12b5608e80f0cc359c9a34178c377bf55e20c980e0ad38d2e6a086de25c"} Dec 06 15:51:23 crc kubenswrapper[4813]: I1206 15:51:23.851282 4813 generic.go:334] "Generic (PLEG): container finished" podID="abeb6c67-8702-4211-9aab-7cdf25c0d078" containerID="b53f8288e238db6836b23603e5c9b94b2e16c5029f6b4d3037214385e94ae11d" exitCode=0 Dec 06 15:51:23 crc kubenswrapper[4813]: I1206 15:51:23.852009 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ws47p" event={"ID":"abeb6c67-8702-4211-9aab-7cdf25c0d078","Type":"ContainerDied","Data":"b53f8288e238db6836b23603e5c9b94b2e16c5029f6b4d3037214385e94ae11d"} Dec 06 15:51:23 crc kubenswrapper[4813]: I1206 15:51:23.852042 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ws47p" event={"ID":"abeb6c67-8702-4211-9aab-7cdf25c0d078","Type":"ContainerStarted","Data":"f33cd0dc63fb253a1e5340ebfb3da977f9b40a187316a22b1e03463ffcc868b8"} Dec 06 15:51:24 crc kubenswrapper[4813]: I1206 15:51:24.572253 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-crqpg"] Dec 06 15:51:24 crc kubenswrapper[4813]: I1206 15:51:24.574216 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-crqpg" Dec 06 15:51:24 crc kubenswrapper[4813]: I1206 15:51:24.577001 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 06 15:51:24 crc kubenswrapper[4813]: I1206 15:51:24.584461 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-crqpg"] Dec 06 15:51:24 crc kubenswrapper[4813]: I1206 15:51:24.737522 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvr5h\" (UniqueName: \"kubernetes.io/projected/4b8fe11e-abdb-42d7-9d88-7d463a808b4b-kube-api-access-tvr5h\") pod \"community-operators-crqpg\" (UID: \"4b8fe11e-abdb-42d7-9d88-7d463a808b4b\") " pod="openshift-marketplace/community-operators-crqpg" Dec 06 15:51:24 crc kubenswrapper[4813]: I1206 15:51:24.737653 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b8fe11e-abdb-42d7-9d88-7d463a808b4b-utilities\") pod \"community-operators-crqpg\" (UID: \"4b8fe11e-abdb-42d7-9d88-7d463a808b4b\") " pod="openshift-marketplace/community-operators-crqpg" Dec 06 15:51:24 crc kubenswrapper[4813]: I1206 15:51:24.737678 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b8fe11e-abdb-42d7-9d88-7d463a808b4b-catalog-content\") pod \"community-operators-crqpg\" (UID: \"4b8fe11e-abdb-42d7-9d88-7d463a808b4b\") " pod="openshift-marketplace/community-operators-crqpg" Dec 06 15:51:24 crc kubenswrapper[4813]: I1206 15:51:24.764326 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-28zj9"] Dec 06 15:51:24 crc kubenswrapper[4813]: I1206 15:51:24.765799 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-28zj9" Dec 06 15:51:24 crc kubenswrapper[4813]: I1206 15:51:24.776412 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 06 15:51:24 crc kubenswrapper[4813]: I1206 15:51:24.801571 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-28zj9"] Dec 06 15:51:24 crc kubenswrapper[4813]: I1206 15:51:24.838631 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46d12276-3ced-4cbe-a060-704f83fcf9a8-catalog-content\") pod \"redhat-operators-28zj9\" (UID: \"46d12276-3ced-4cbe-a060-704f83fcf9a8\") " pod="openshift-marketplace/redhat-operators-28zj9" Dec 06 15:51:24 crc kubenswrapper[4813]: I1206 15:51:24.838989 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvr5h\" (UniqueName: \"kubernetes.io/projected/4b8fe11e-abdb-42d7-9d88-7d463a808b4b-kube-api-access-tvr5h\") pod \"community-operators-crqpg\" (UID: \"4b8fe11e-abdb-42d7-9d88-7d463a808b4b\") " pod="openshift-marketplace/community-operators-crqpg" Dec 06 15:51:24 crc kubenswrapper[4813]: I1206 15:51:24.839125 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46d12276-3ced-4cbe-a060-704f83fcf9a8-utilities\") pod \"redhat-operators-28zj9\" (UID: \"46d12276-3ced-4cbe-a060-704f83fcf9a8\") " pod="openshift-marketplace/redhat-operators-28zj9" Dec 06 15:51:24 crc kubenswrapper[4813]: I1206 15:51:24.839237 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kqz2\" (UniqueName: \"kubernetes.io/projected/46d12276-3ced-4cbe-a060-704f83fcf9a8-kube-api-access-5kqz2\") pod \"redhat-operators-28zj9\" (UID: \"46d12276-3ced-4cbe-a060-704f83fcf9a8\") " pod="openshift-marketplace/redhat-operators-28zj9" Dec 06 15:51:24 crc kubenswrapper[4813]: I1206 15:51:24.839463 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b8fe11e-abdb-42d7-9d88-7d463a808b4b-utilities\") pod \"community-operators-crqpg\" (UID: \"4b8fe11e-abdb-42d7-9d88-7d463a808b4b\") " pod="openshift-marketplace/community-operators-crqpg" Dec 06 15:51:24 crc kubenswrapper[4813]: I1206 15:51:24.839543 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b8fe11e-abdb-42d7-9d88-7d463a808b4b-catalog-content\") pod \"community-operators-crqpg\" (UID: \"4b8fe11e-abdb-42d7-9d88-7d463a808b4b\") " pod="openshift-marketplace/community-operators-crqpg" Dec 06 15:51:24 crc kubenswrapper[4813]: I1206 15:51:24.840111 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b8fe11e-abdb-42d7-9d88-7d463a808b4b-catalog-content\") pod \"community-operators-crqpg\" (UID: \"4b8fe11e-abdb-42d7-9d88-7d463a808b4b\") " pod="openshift-marketplace/community-operators-crqpg" Dec 06 15:51:24 crc kubenswrapper[4813]: I1206 15:51:24.840344 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b8fe11e-abdb-42d7-9d88-7d463a808b4b-utilities\") pod \"community-operators-crqpg\" (UID: \"4b8fe11e-abdb-42d7-9d88-7d463a808b4b\") " pod="openshift-marketplace/community-operators-crqpg" Dec 06 15:51:24 crc kubenswrapper[4813]: I1206 15:51:24.858105 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xwq7l" event={"ID":"ca46ce83-a423-478a-9d44-d45a0ef28ef3","Type":"ContainerStarted","Data":"428db26c20a6c2e0b327b8c6dde6fd68b0da2f34caa89524f5a40c6bf6170eb1"} Dec 06 15:51:24 crc kubenswrapper[4813]: I1206 15:51:24.860002 4813 generic.go:334] "Generic (PLEG): container finished" podID="abeb6c67-8702-4211-9aab-7cdf25c0d078" containerID="f5b6a8c952eb7236e94d237407787390877d4ac9a4de7b654bed2d743c302b20" exitCode=0 Dec 06 15:51:24 crc kubenswrapper[4813]: I1206 15:51:24.860111 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ws47p" event={"ID":"abeb6c67-8702-4211-9aab-7cdf25c0d078","Type":"ContainerDied","Data":"f5b6a8c952eb7236e94d237407787390877d4ac9a4de7b654bed2d743c302b20"} Dec 06 15:51:24 crc kubenswrapper[4813]: I1206 15:51:24.879273 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvr5h\" (UniqueName: \"kubernetes.io/projected/4b8fe11e-abdb-42d7-9d88-7d463a808b4b-kube-api-access-tvr5h\") pod \"community-operators-crqpg\" (UID: \"4b8fe11e-abdb-42d7-9d88-7d463a808b4b\") " pod="openshift-marketplace/community-operators-crqpg" Dec 06 15:51:24 crc kubenswrapper[4813]: I1206 15:51:24.893544 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-crqpg" Dec 06 15:51:24 crc kubenswrapper[4813]: I1206 15:51:24.940162 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46d12276-3ced-4cbe-a060-704f83fcf9a8-catalog-content\") pod \"redhat-operators-28zj9\" (UID: \"46d12276-3ced-4cbe-a060-704f83fcf9a8\") " pod="openshift-marketplace/redhat-operators-28zj9" Dec 06 15:51:24 crc kubenswrapper[4813]: I1206 15:51:24.940277 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46d12276-3ced-4cbe-a060-704f83fcf9a8-utilities\") pod \"redhat-operators-28zj9\" (UID: \"46d12276-3ced-4cbe-a060-704f83fcf9a8\") " pod="openshift-marketplace/redhat-operators-28zj9" Dec 06 15:51:24 crc kubenswrapper[4813]: I1206 15:51:24.940301 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kqz2\" (UniqueName: \"kubernetes.io/projected/46d12276-3ced-4cbe-a060-704f83fcf9a8-kube-api-access-5kqz2\") pod \"redhat-operators-28zj9\" (UID: \"46d12276-3ced-4cbe-a060-704f83fcf9a8\") " pod="openshift-marketplace/redhat-operators-28zj9" Dec 06 15:51:24 crc kubenswrapper[4813]: I1206 15:51:24.940841 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46d12276-3ced-4cbe-a060-704f83fcf9a8-catalog-content\") pod \"redhat-operators-28zj9\" (UID: \"46d12276-3ced-4cbe-a060-704f83fcf9a8\") " pod="openshift-marketplace/redhat-operators-28zj9" Dec 06 15:51:24 crc kubenswrapper[4813]: I1206 15:51:24.941068 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46d12276-3ced-4cbe-a060-704f83fcf9a8-utilities\") pod \"redhat-operators-28zj9\" (UID: \"46d12276-3ced-4cbe-a060-704f83fcf9a8\") " pod="openshift-marketplace/redhat-operators-28zj9" Dec 06 15:51:24 crc kubenswrapper[4813]: I1206 15:51:24.963347 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kqz2\" (UniqueName: \"kubernetes.io/projected/46d12276-3ced-4cbe-a060-704f83fcf9a8-kube-api-access-5kqz2\") pod \"redhat-operators-28zj9\" (UID: \"46d12276-3ced-4cbe-a060-704f83fcf9a8\") " pod="openshift-marketplace/redhat-operators-28zj9" Dec 06 15:51:25 crc kubenswrapper[4813]: I1206 15:51:25.080776 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-28zj9" Dec 06 15:51:25 crc kubenswrapper[4813]: I1206 15:51:25.333793 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-crqpg"] Dec 06 15:51:25 crc kubenswrapper[4813]: W1206 15:51:25.343072 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b8fe11e_abdb_42d7_9d88_7d463a808b4b.slice/crio-6195f8ae989b989f17349a2150425eed19b2caa99a28d1f1decc6deb6fda8613 WatchSource:0}: Error finding container 6195f8ae989b989f17349a2150425eed19b2caa99a28d1f1decc6deb6fda8613: Status 404 returned error can't find the container with id 6195f8ae989b989f17349a2150425eed19b2caa99a28d1f1decc6deb6fda8613 Dec 06 15:51:25 crc kubenswrapper[4813]: I1206 15:51:25.347464 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-28zj9"] Dec 06 15:51:25 crc kubenswrapper[4813]: W1206 15:51:25.359409 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46d12276_3ced_4cbe_a060_704f83fcf9a8.slice/crio-713c7112880056eaa34e7314e2925a192edc0762fdcd7528a941304461fb113e WatchSource:0}: Error finding container 713c7112880056eaa34e7314e2925a192edc0762fdcd7528a941304461fb113e: Status 404 returned error can't find the container with id 713c7112880056eaa34e7314e2925a192edc0762fdcd7528a941304461fb113e Dec 06 15:51:25 crc kubenswrapper[4813]: I1206 15:51:25.866725 4813 generic.go:334] "Generic (PLEG): container finished" podID="ca46ce83-a423-478a-9d44-d45a0ef28ef3" containerID="428db26c20a6c2e0b327b8c6dde6fd68b0da2f34caa89524f5a40c6bf6170eb1" exitCode=0 Dec 06 15:51:25 crc kubenswrapper[4813]: I1206 15:51:25.866986 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xwq7l" event={"ID":"ca46ce83-a423-478a-9d44-d45a0ef28ef3","Type":"ContainerDied","Data":"428db26c20a6c2e0b327b8c6dde6fd68b0da2f34caa89524f5a40c6bf6170eb1"} Dec 06 15:51:25 crc kubenswrapper[4813]: I1206 15:51:25.869894 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ws47p" event={"ID":"abeb6c67-8702-4211-9aab-7cdf25c0d078","Type":"ContainerStarted","Data":"5895c502e53961f1d29f13383b26dbfa31d5554bc6ed2a71f6415bb788d68c5e"} Dec 06 15:51:25 crc kubenswrapper[4813]: I1206 15:51:25.871215 4813 generic.go:334] "Generic (PLEG): container finished" podID="46d12276-3ced-4cbe-a060-704f83fcf9a8" containerID="ba3404087534aa293bcba78675b7572676867b4657586b8a2bf88968b94fde63" exitCode=0 Dec 06 15:51:25 crc kubenswrapper[4813]: I1206 15:51:25.871311 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-28zj9" event={"ID":"46d12276-3ced-4cbe-a060-704f83fcf9a8","Type":"ContainerDied","Data":"ba3404087534aa293bcba78675b7572676867b4657586b8a2bf88968b94fde63"} Dec 06 15:51:25 crc kubenswrapper[4813]: I1206 15:51:25.871365 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-28zj9" event={"ID":"46d12276-3ced-4cbe-a060-704f83fcf9a8","Type":"ContainerStarted","Data":"713c7112880056eaa34e7314e2925a192edc0762fdcd7528a941304461fb113e"} Dec 06 15:51:25 crc kubenswrapper[4813]: I1206 15:51:25.873006 4813 generic.go:334] "Generic (PLEG): container finished" podID="4b8fe11e-abdb-42d7-9d88-7d463a808b4b" containerID="51e2a084c8beef82d4e4a635209ba084466da3cdc81d116f580ec20fa5300ed9" exitCode=0 Dec 06 15:51:25 crc kubenswrapper[4813]: I1206 15:51:25.873028 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-crqpg" event={"ID":"4b8fe11e-abdb-42d7-9d88-7d463a808b4b","Type":"ContainerDied","Data":"51e2a084c8beef82d4e4a635209ba084466da3cdc81d116f580ec20fa5300ed9"} Dec 06 15:51:25 crc kubenswrapper[4813]: I1206 15:51:25.873312 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-crqpg" event={"ID":"4b8fe11e-abdb-42d7-9d88-7d463a808b4b","Type":"ContainerStarted","Data":"6195f8ae989b989f17349a2150425eed19b2caa99a28d1f1decc6deb6fda8613"} Dec 06 15:51:25 crc kubenswrapper[4813]: I1206 15:51:25.924787 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ws47p" podStartSLOduration=2.5154568299999998 podStartE2EDuration="3.92476589s" podCreationTimestamp="2025-12-06 15:51:22 +0000 UTC" firstStartedPulling="2025-12-06 15:51:23.854091489 +0000 UTC m=+323.744971065" lastFinishedPulling="2025-12-06 15:51:25.263400549 +0000 UTC m=+325.154280125" observedRunningTime="2025-12-06 15:51:25.923881356 +0000 UTC m=+325.814760932" watchObservedRunningTime="2025-12-06 15:51:25.92476589 +0000 UTC m=+325.815645476" Dec 06 15:51:26 crc kubenswrapper[4813]: I1206 15:51:26.879375 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-crqpg" event={"ID":"4b8fe11e-abdb-42d7-9d88-7d463a808b4b","Type":"ContainerStarted","Data":"b647a75da487d44c75a12be9e51718495a32ac68edb97d6c90ff00493aad6dc9"} Dec 06 15:51:26 crc kubenswrapper[4813]: I1206 15:51:26.881889 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xwq7l" event={"ID":"ca46ce83-a423-478a-9d44-d45a0ef28ef3","Type":"ContainerStarted","Data":"b2d3b2a1cd1bd6e4311d23a6834170b825f4e94915c8043a52bde891725ceae7"} Dec 06 15:51:26 crc kubenswrapper[4813]: I1206 15:51:26.883221 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-28zj9" event={"ID":"46d12276-3ced-4cbe-a060-704f83fcf9a8","Type":"ContainerStarted","Data":"1005adb9b645de31a31dffe7555f7cd1e93d821f22167571936925b9bee04ed7"} Dec 06 15:51:26 crc kubenswrapper[4813]: I1206 15:51:26.928030 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xwq7l" podStartSLOduration=3.557203337 podStartE2EDuration="5.928005711s" podCreationTimestamp="2025-12-06 15:51:21 +0000 UTC" firstStartedPulling="2025-12-06 15:51:23.850500451 +0000 UTC m=+323.741380027" lastFinishedPulling="2025-12-06 15:51:26.221302815 +0000 UTC m=+326.112182401" observedRunningTime="2025-12-06 15:51:26.927637291 +0000 UTC m=+326.818516867" watchObservedRunningTime="2025-12-06 15:51:26.928005711 +0000 UTC m=+326.818885287" Dec 06 15:51:27 crc kubenswrapper[4813]: I1206 15:51:27.891432 4813 generic.go:334] "Generic (PLEG): container finished" podID="4b8fe11e-abdb-42d7-9d88-7d463a808b4b" containerID="b647a75da487d44c75a12be9e51718495a32ac68edb97d6c90ff00493aad6dc9" exitCode=0 Dec 06 15:51:27 crc kubenswrapper[4813]: I1206 15:51:27.891493 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-crqpg" event={"ID":"4b8fe11e-abdb-42d7-9d88-7d463a808b4b","Type":"ContainerDied","Data":"b647a75da487d44c75a12be9e51718495a32ac68edb97d6c90ff00493aad6dc9"} Dec 06 15:51:27 crc kubenswrapper[4813]: I1206 15:51:27.895142 4813 generic.go:334] "Generic (PLEG): container finished" podID="46d12276-3ced-4cbe-a060-704f83fcf9a8" containerID="1005adb9b645de31a31dffe7555f7cd1e93d821f22167571936925b9bee04ed7" exitCode=0 Dec 06 15:51:27 crc kubenswrapper[4813]: I1206 15:51:27.895199 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-28zj9" event={"ID":"46d12276-3ced-4cbe-a060-704f83fcf9a8","Type":"ContainerDied","Data":"1005adb9b645de31a31dffe7555f7cd1e93d821f22167571936925b9bee04ed7"} Dec 06 15:51:28 crc kubenswrapper[4813]: I1206 15:51:28.901993 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-crqpg" event={"ID":"4b8fe11e-abdb-42d7-9d88-7d463a808b4b","Type":"ContainerStarted","Data":"e5de40fdbe3e1a617f833e1beb07c4c052db431a19dd935c2a813fb94440d9e1"} Dec 06 15:51:28 crc kubenswrapper[4813]: I1206 15:51:28.921720 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-28zj9" event={"ID":"46d12276-3ced-4cbe-a060-704f83fcf9a8","Type":"ContainerStarted","Data":"4ef82a0609a2c91ccd93f8c0829f28f7b1ca8476fdd69fe0fcfadd87be45df05"} Dec 06 15:51:28 crc kubenswrapper[4813]: I1206 15:51:28.939811 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-crqpg" podStartSLOduration=2.528498384 podStartE2EDuration="4.939797552s" podCreationTimestamp="2025-12-06 15:51:24 +0000 UTC" firstStartedPulling="2025-12-06 15:51:25.874796449 +0000 UTC m=+325.765676035" lastFinishedPulling="2025-12-06 15:51:28.286095607 +0000 UTC m=+328.176975203" observedRunningTime="2025-12-06 15:51:28.938924589 +0000 UTC m=+328.829804165" watchObservedRunningTime="2025-12-06 15:51:28.939797552 +0000 UTC m=+328.830677128" Dec 06 15:51:28 crc kubenswrapper[4813]: I1206 15:51:28.962499 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-28zj9" podStartSLOduration=2.54264428 podStartE2EDuration="4.962482291s" podCreationTimestamp="2025-12-06 15:51:24 +0000 UTC" firstStartedPulling="2025-12-06 15:51:25.872339012 +0000 UTC m=+325.763218588" lastFinishedPulling="2025-12-06 15:51:28.292177013 +0000 UTC m=+328.183056599" observedRunningTime="2025-12-06 15:51:28.962437389 +0000 UTC m=+328.853316965" watchObservedRunningTime="2025-12-06 15:51:28.962482291 +0000 UTC m=+328.853361877" Dec 06 15:51:29 crc kubenswrapper[4813]: I1206 15:51:29.962411 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6bbs5"] Dec 06 15:51:29 crc kubenswrapper[4813]: I1206 15:51:29.962777 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-6bbs5" podUID="1d5678dc-7a74-4fff-818c-e34321037686" containerName="controller-manager" containerID="cri-o://fd283e9e9706af8249ef143f4974d9053373dd945b4e3593bec104acbcc002c2" gracePeriod=30 Dec 06 15:51:29 crc kubenswrapper[4813]: I1206 15:51:29.969867 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q"] Dec 06 15:51:29 crc kubenswrapper[4813]: I1206 15:51:29.970057 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q" podUID="4b409627-4716-4bdd-a660-4f88693150a2" containerName="route-controller-manager" containerID="cri-o://2f140a4a1f10630431e04b37c794615c8c1ed29d7b4d0ce308142240350cea09" gracePeriod=30 Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.576397 4813 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-6bbs5 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.576469 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-6bbs5" podUID="1d5678dc-7a74-4fff-818c-e34321037686" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.737113 4813 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-6896q container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.737376 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q" podUID="4b409627-4716-4bdd-a660-4f88693150a2" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.834763 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-6bbs5" Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.920093 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lkhl\" (UniqueName: \"kubernetes.io/projected/1d5678dc-7a74-4fff-818c-e34321037686-kube-api-access-8lkhl\") pod \"1d5678dc-7a74-4fff-818c-e34321037686\" (UID: \"1d5678dc-7a74-4fff-818c-e34321037686\") " Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.920156 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1d5678dc-7a74-4fff-818c-e34321037686-client-ca\") pod \"1d5678dc-7a74-4fff-818c-e34321037686\" (UID: \"1d5678dc-7a74-4fff-818c-e34321037686\") " Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.920204 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1d5678dc-7a74-4fff-818c-e34321037686-proxy-ca-bundles\") pod \"1d5678dc-7a74-4fff-818c-e34321037686\" (UID: \"1d5678dc-7a74-4fff-818c-e34321037686\") " Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.920944 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d5678dc-7a74-4fff-818c-e34321037686-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "1d5678dc-7a74-4fff-818c-e34321037686" (UID: "1d5678dc-7a74-4fff-818c-e34321037686"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.921506 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d5678dc-7a74-4fff-818c-e34321037686-client-ca" (OuterVolumeSpecName: "client-ca") pod "1d5678dc-7a74-4fff-818c-e34321037686" (UID: "1d5678dc-7a74-4fff-818c-e34321037686"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.921620 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d5678dc-7a74-4fff-818c-e34321037686-config\") pod \"1d5678dc-7a74-4fff-818c-e34321037686\" (UID: \"1d5678dc-7a74-4fff-818c-e34321037686\") " Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.922154 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d5678dc-7a74-4fff-818c-e34321037686-config" (OuterVolumeSpecName: "config") pod "1d5678dc-7a74-4fff-818c-e34321037686" (UID: "1d5678dc-7a74-4fff-818c-e34321037686"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.922451 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d5678dc-7a74-4fff-818c-e34321037686-serving-cert\") pod \"1d5678dc-7a74-4fff-818c-e34321037686\" (UID: \"1d5678dc-7a74-4fff-818c-e34321037686\") " Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.923109 4813 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1d5678dc-7a74-4fff-818c-e34321037686-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.923129 4813 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1d5678dc-7a74-4fff-818c-e34321037686-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.923139 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d5678dc-7a74-4fff-818c-e34321037686-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.925407 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d5678dc-7a74-4fff-818c-e34321037686-kube-api-access-8lkhl" (OuterVolumeSpecName: "kube-api-access-8lkhl") pod "1d5678dc-7a74-4fff-818c-e34321037686" (UID: "1d5678dc-7a74-4fff-818c-e34321037686"). InnerVolumeSpecName "kube-api-access-8lkhl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.925570 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d5678dc-7a74-4fff-818c-e34321037686-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1d5678dc-7a74-4fff-818c-e34321037686" (UID: "1d5678dc-7a74-4fff-818c-e34321037686"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.929348 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q" Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.933564 4813 generic.go:334] "Generic (PLEG): container finished" podID="4b409627-4716-4bdd-a660-4f88693150a2" containerID="2f140a4a1f10630431e04b37c794615c8c1ed29d7b4d0ce308142240350cea09" exitCode=0 Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.934838 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q" event={"ID":"4b409627-4716-4bdd-a660-4f88693150a2","Type":"ContainerDied","Data":"2f140a4a1f10630431e04b37c794615c8c1ed29d7b4d0ce308142240350cea09"} Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.934882 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q" event={"ID":"4b409627-4716-4bdd-a660-4f88693150a2","Type":"ContainerDied","Data":"2c79dfb191db5625286c5f1e6423c158411cee10845f2895cbd98f97f526f35f"} Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.934899 4813 scope.go:117] "RemoveContainer" containerID="2f140a4a1f10630431e04b37c794615c8c1ed29d7b4d0ce308142240350cea09" Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.935017 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q" Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.937138 4813 generic.go:334] "Generic (PLEG): container finished" podID="1d5678dc-7a74-4fff-818c-e34321037686" containerID="fd283e9e9706af8249ef143f4974d9053373dd945b4e3593bec104acbcc002c2" exitCode=0 Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.937184 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-6bbs5" event={"ID":"1d5678dc-7a74-4fff-818c-e34321037686","Type":"ContainerDied","Data":"fd283e9e9706af8249ef143f4974d9053373dd945b4e3593bec104acbcc002c2"} Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.937213 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-6bbs5" event={"ID":"1d5678dc-7a74-4fff-818c-e34321037686","Type":"ContainerDied","Data":"beffd2960d07cdfd8fb916176928a691fd0c55570d38ae8bb6ef245ce63da5cc"} Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.937282 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-6bbs5" Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.961632 4813 scope.go:117] "RemoveContainer" containerID="2f140a4a1f10630431e04b37c794615c8c1ed29d7b4d0ce308142240350cea09" Dec 06 15:51:30 crc kubenswrapper[4813]: E1206 15:51:30.963669 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f140a4a1f10630431e04b37c794615c8c1ed29d7b4d0ce308142240350cea09\": container with ID starting with 2f140a4a1f10630431e04b37c794615c8c1ed29d7b4d0ce308142240350cea09 not found: ID does not exist" containerID="2f140a4a1f10630431e04b37c794615c8c1ed29d7b4d0ce308142240350cea09" Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.963698 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f140a4a1f10630431e04b37c794615c8c1ed29d7b4d0ce308142240350cea09"} err="failed to get container status \"2f140a4a1f10630431e04b37c794615c8c1ed29d7b4d0ce308142240350cea09\": rpc error: code = NotFound desc = could not find container \"2f140a4a1f10630431e04b37c794615c8c1ed29d7b4d0ce308142240350cea09\": container with ID starting with 2f140a4a1f10630431e04b37c794615c8c1ed29d7b4d0ce308142240350cea09 not found: ID does not exist" Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.963717 4813 scope.go:117] "RemoveContainer" containerID="fd283e9e9706af8249ef143f4974d9053373dd945b4e3593bec104acbcc002c2" Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.983357 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6bbs5"] Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.983403 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6bbs5"] Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.984197 4813 scope.go:117] "RemoveContainer" containerID="fd283e9e9706af8249ef143f4974d9053373dd945b4e3593bec104acbcc002c2" Dec 06 15:51:30 crc kubenswrapper[4813]: E1206 15:51:30.984764 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd283e9e9706af8249ef143f4974d9053373dd945b4e3593bec104acbcc002c2\": container with ID starting with fd283e9e9706af8249ef143f4974d9053373dd945b4e3593bec104acbcc002c2 not found: ID does not exist" containerID="fd283e9e9706af8249ef143f4974d9053373dd945b4e3593bec104acbcc002c2" Dec 06 15:51:30 crc kubenswrapper[4813]: I1206 15:51:30.984787 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd283e9e9706af8249ef143f4974d9053373dd945b4e3593bec104acbcc002c2"} err="failed to get container status \"fd283e9e9706af8249ef143f4974d9053373dd945b4e3593bec104acbcc002c2\": rpc error: code = NotFound desc = could not find container \"fd283e9e9706af8249ef143f4974d9053373dd945b4e3593bec104acbcc002c2\": container with ID starting with fd283e9e9706af8249ef143f4974d9053373dd945b4e3593bec104acbcc002c2 not found: ID does not exist" Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.023641 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mm2f\" (UniqueName: \"kubernetes.io/projected/4b409627-4716-4bdd-a660-4f88693150a2-kube-api-access-7mm2f\") pod \"4b409627-4716-4bdd-a660-4f88693150a2\" (UID: \"4b409627-4716-4bdd-a660-4f88693150a2\") " Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.023722 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b409627-4716-4bdd-a660-4f88693150a2-serving-cert\") pod \"4b409627-4716-4bdd-a660-4f88693150a2\" (UID: \"4b409627-4716-4bdd-a660-4f88693150a2\") " Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.023813 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b409627-4716-4bdd-a660-4f88693150a2-config\") pod \"4b409627-4716-4bdd-a660-4f88693150a2\" (UID: \"4b409627-4716-4bdd-a660-4f88693150a2\") " Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.023831 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4b409627-4716-4bdd-a660-4f88693150a2-client-ca\") pod \"4b409627-4716-4bdd-a660-4f88693150a2\" (UID: \"4b409627-4716-4bdd-a660-4f88693150a2\") " Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.023998 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d5678dc-7a74-4fff-818c-e34321037686-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.024013 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lkhl\" (UniqueName: \"kubernetes.io/projected/1d5678dc-7a74-4fff-818c-e34321037686-kube-api-access-8lkhl\") on node \"crc\" DevicePath \"\"" Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.024425 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b409627-4716-4bdd-a660-4f88693150a2-client-ca" (OuterVolumeSpecName: "client-ca") pod "4b409627-4716-4bdd-a660-4f88693150a2" (UID: "4b409627-4716-4bdd-a660-4f88693150a2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.024752 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b409627-4716-4bdd-a660-4f88693150a2-config" (OuterVolumeSpecName: "config") pod "4b409627-4716-4bdd-a660-4f88693150a2" (UID: "4b409627-4716-4bdd-a660-4f88693150a2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.026868 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b409627-4716-4bdd-a660-4f88693150a2-kube-api-access-7mm2f" (OuterVolumeSpecName: "kube-api-access-7mm2f") pod "4b409627-4716-4bdd-a660-4f88693150a2" (UID: "4b409627-4716-4bdd-a660-4f88693150a2"). InnerVolumeSpecName "kube-api-access-7mm2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.028560 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b409627-4716-4bdd-a660-4f88693150a2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4b409627-4716-4bdd-a660-4f88693150a2" (UID: "4b409627-4716-4bdd-a660-4f88693150a2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.076817 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-767f84-ktpsj"] Dec 06 15:51:31 crc kubenswrapper[4813]: E1206 15:51:31.076996 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b409627-4716-4bdd-a660-4f88693150a2" containerName="route-controller-manager" Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.077007 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b409627-4716-4bdd-a660-4f88693150a2" containerName="route-controller-manager" Dec 06 15:51:31 crc kubenswrapper[4813]: E1206 15:51:31.077020 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d5678dc-7a74-4fff-818c-e34321037686" containerName="controller-manager" Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.077026 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d5678dc-7a74-4fff-818c-e34321037686" containerName="controller-manager" Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.077106 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d5678dc-7a74-4fff-818c-e34321037686" containerName="controller-manager" Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.077119 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b409627-4716-4bdd-a660-4f88693150a2" containerName="route-controller-manager" Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.077441 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-767f84-ktpsj" Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.088973 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-767f84-ktpsj"] Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.125503 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b409627-4716-4bdd-a660-4f88693150a2-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.125529 4813 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4b409627-4716-4bdd-a660-4f88693150a2-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.125541 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mm2f\" (UniqueName: \"kubernetes.io/projected/4b409627-4716-4bdd-a660-4f88693150a2-kube-api-access-7mm2f\") on node \"crc\" DevicePath \"\"" Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.125549 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b409627-4716-4bdd-a660-4f88693150a2-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.226539 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8xkb\" (UniqueName: \"kubernetes.io/projected/3a19d1ed-a1c3-4060-a254-a29eefa188bc-kube-api-access-z8xkb\") pod \"route-controller-manager-767f84-ktpsj\" (UID: \"3a19d1ed-a1c3-4060-a254-a29eefa188bc\") " pod="openshift-route-controller-manager/route-controller-manager-767f84-ktpsj" Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.226601 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a19d1ed-a1c3-4060-a254-a29eefa188bc-config\") pod \"route-controller-manager-767f84-ktpsj\" (UID: \"3a19d1ed-a1c3-4060-a254-a29eefa188bc\") " pod="openshift-route-controller-manager/route-controller-manager-767f84-ktpsj" Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.226626 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3a19d1ed-a1c3-4060-a254-a29eefa188bc-client-ca\") pod \"route-controller-manager-767f84-ktpsj\" (UID: \"3a19d1ed-a1c3-4060-a254-a29eefa188bc\") " pod="openshift-route-controller-manager/route-controller-manager-767f84-ktpsj" Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.226690 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a19d1ed-a1c3-4060-a254-a29eefa188bc-serving-cert\") pod \"route-controller-manager-767f84-ktpsj\" (UID: \"3a19d1ed-a1c3-4060-a254-a29eefa188bc\") " pod="openshift-route-controller-manager/route-controller-manager-767f84-ktpsj" Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.258721 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q"] Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.265329 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6896q"] Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.328192 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a19d1ed-a1c3-4060-a254-a29eefa188bc-serving-cert\") pod \"route-controller-manager-767f84-ktpsj\" (UID: \"3a19d1ed-a1c3-4060-a254-a29eefa188bc\") " pod="openshift-route-controller-manager/route-controller-manager-767f84-ktpsj" Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.328600 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8xkb\" (UniqueName: \"kubernetes.io/projected/3a19d1ed-a1c3-4060-a254-a29eefa188bc-kube-api-access-z8xkb\") pod \"route-controller-manager-767f84-ktpsj\" (UID: \"3a19d1ed-a1c3-4060-a254-a29eefa188bc\") " pod="openshift-route-controller-manager/route-controller-manager-767f84-ktpsj" Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.328750 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a19d1ed-a1c3-4060-a254-a29eefa188bc-config\") pod \"route-controller-manager-767f84-ktpsj\" (UID: \"3a19d1ed-a1c3-4060-a254-a29eefa188bc\") " pod="openshift-route-controller-manager/route-controller-manager-767f84-ktpsj" Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.328951 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3a19d1ed-a1c3-4060-a254-a29eefa188bc-client-ca\") pod \"route-controller-manager-767f84-ktpsj\" (UID: \"3a19d1ed-a1c3-4060-a254-a29eefa188bc\") " pod="openshift-route-controller-manager/route-controller-manager-767f84-ktpsj" Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.330200 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3a19d1ed-a1c3-4060-a254-a29eefa188bc-client-ca\") pod \"route-controller-manager-767f84-ktpsj\" (UID: \"3a19d1ed-a1c3-4060-a254-a29eefa188bc\") " pod="openshift-route-controller-manager/route-controller-manager-767f84-ktpsj" Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.330328 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a19d1ed-a1c3-4060-a254-a29eefa188bc-config\") pod \"route-controller-manager-767f84-ktpsj\" (UID: \"3a19d1ed-a1c3-4060-a254-a29eefa188bc\") " pod="openshift-route-controller-manager/route-controller-manager-767f84-ktpsj" Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.331751 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a19d1ed-a1c3-4060-a254-a29eefa188bc-serving-cert\") pod \"route-controller-manager-767f84-ktpsj\" (UID: \"3a19d1ed-a1c3-4060-a254-a29eefa188bc\") " pod="openshift-route-controller-manager/route-controller-manager-767f84-ktpsj" Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.365137 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8xkb\" (UniqueName: \"kubernetes.io/projected/3a19d1ed-a1c3-4060-a254-a29eefa188bc-kube-api-access-z8xkb\") pod \"route-controller-manager-767f84-ktpsj\" (UID: \"3a19d1ed-a1c3-4060-a254-a29eefa188bc\") " pod="openshift-route-controller-manager/route-controller-manager-767f84-ktpsj" Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.396022 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-767f84-ktpsj" Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.624920 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-767f84-ktpsj"] Dec 06 15:51:31 crc kubenswrapper[4813]: I1206 15:51:31.959119 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-767f84-ktpsj" event={"ID":"3a19d1ed-a1c3-4060-a254-a29eefa188bc","Type":"ContainerStarted","Data":"1b3498f5b85409739fbce7d4e8aefb1bc80ea636ce8c280f710fe7b3139d2e85"} Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.079867 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-796947dbf8-npktj"] Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.081752 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-796947dbf8-npktj" Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.084771 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.084777 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.084920 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.085095 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.085542 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.085617 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.098229 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.111971 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-796947dbf8-npktj"] Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.245085 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cf16c324-2ece-43f9-8822-2e7ce9f06e7a-client-ca\") pod \"controller-manager-796947dbf8-npktj\" (UID: \"cf16c324-2ece-43f9-8822-2e7ce9f06e7a\") " pod="openshift-controller-manager/controller-manager-796947dbf8-npktj" Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.245132 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgtx4\" (UniqueName: \"kubernetes.io/projected/cf16c324-2ece-43f9-8822-2e7ce9f06e7a-kube-api-access-dgtx4\") pod \"controller-manager-796947dbf8-npktj\" (UID: \"cf16c324-2ece-43f9-8822-2e7ce9f06e7a\") " pod="openshift-controller-manager/controller-manager-796947dbf8-npktj" Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.245153 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cf16c324-2ece-43f9-8822-2e7ce9f06e7a-proxy-ca-bundles\") pod \"controller-manager-796947dbf8-npktj\" (UID: \"cf16c324-2ece-43f9-8822-2e7ce9f06e7a\") " pod="openshift-controller-manager/controller-manager-796947dbf8-npktj" Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.245197 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf16c324-2ece-43f9-8822-2e7ce9f06e7a-serving-cert\") pod \"controller-manager-796947dbf8-npktj\" (UID: \"cf16c324-2ece-43f9-8822-2e7ce9f06e7a\") " pod="openshift-controller-manager/controller-manager-796947dbf8-npktj" Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.245227 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf16c324-2ece-43f9-8822-2e7ce9f06e7a-config\") pod \"controller-manager-796947dbf8-npktj\" (UID: \"cf16c324-2ece-43f9-8822-2e7ce9f06e7a\") " pod="openshift-controller-manager/controller-manager-796947dbf8-npktj" Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.306168 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xwq7l" Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.306299 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xwq7l" Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.347897 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cf16c324-2ece-43f9-8822-2e7ce9f06e7a-proxy-ca-bundles\") pod \"controller-manager-796947dbf8-npktj\" (UID: \"cf16c324-2ece-43f9-8822-2e7ce9f06e7a\") " pod="openshift-controller-manager/controller-manager-796947dbf8-npktj" Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.348125 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf16c324-2ece-43f9-8822-2e7ce9f06e7a-serving-cert\") pod \"controller-manager-796947dbf8-npktj\" (UID: \"cf16c324-2ece-43f9-8822-2e7ce9f06e7a\") " pod="openshift-controller-manager/controller-manager-796947dbf8-npktj" Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.348360 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf16c324-2ece-43f9-8822-2e7ce9f06e7a-config\") pod \"controller-manager-796947dbf8-npktj\" (UID: \"cf16c324-2ece-43f9-8822-2e7ce9f06e7a\") " pod="openshift-controller-manager/controller-manager-796947dbf8-npktj" Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.348465 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cf16c324-2ece-43f9-8822-2e7ce9f06e7a-client-ca\") pod \"controller-manager-796947dbf8-npktj\" (UID: \"cf16c324-2ece-43f9-8822-2e7ce9f06e7a\") " pod="openshift-controller-manager/controller-manager-796947dbf8-npktj" Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.348563 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgtx4\" (UniqueName: \"kubernetes.io/projected/cf16c324-2ece-43f9-8822-2e7ce9f06e7a-kube-api-access-dgtx4\") pod \"controller-manager-796947dbf8-npktj\" (UID: \"cf16c324-2ece-43f9-8822-2e7ce9f06e7a\") " pod="openshift-controller-manager/controller-manager-796947dbf8-npktj" Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.352008 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cf16c324-2ece-43f9-8822-2e7ce9f06e7a-proxy-ca-bundles\") pod \"controller-manager-796947dbf8-npktj\" (UID: \"cf16c324-2ece-43f9-8822-2e7ce9f06e7a\") " pod="openshift-controller-manager/controller-manager-796947dbf8-npktj" Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.356075 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf16c324-2ece-43f9-8822-2e7ce9f06e7a-config\") pod \"controller-manager-796947dbf8-npktj\" (UID: \"cf16c324-2ece-43f9-8822-2e7ce9f06e7a\") " pod="openshift-controller-manager/controller-manager-796947dbf8-npktj" Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.357232 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cf16c324-2ece-43f9-8822-2e7ce9f06e7a-client-ca\") pod \"controller-manager-796947dbf8-npktj\" (UID: \"cf16c324-2ece-43f9-8822-2e7ce9f06e7a\") " pod="openshift-controller-manager/controller-manager-796947dbf8-npktj" Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.360800 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf16c324-2ece-43f9-8822-2e7ce9f06e7a-serving-cert\") pod \"controller-manager-796947dbf8-npktj\" (UID: \"cf16c324-2ece-43f9-8822-2e7ce9f06e7a\") " pod="openshift-controller-manager/controller-manager-796947dbf8-npktj" Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.368649 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgtx4\" (UniqueName: \"kubernetes.io/projected/cf16c324-2ece-43f9-8822-2e7ce9f06e7a-kube-api-access-dgtx4\") pod \"controller-manager-796947dbf8-npktj\" (UID: \"cf16c324-2ece-43f9-8822-2e7ce9f06e7a\") " pod="openshift-controller-manager/controller-manager-796947dbf8-npktj" Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.409021 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xwq7l" Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.415470 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-796947dbf8-npktj" Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.496339 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d5678dc-7a74-4fff-818c-e34321037686" path="/var/lib/kubelet/pods/1d5678dc-7a74-4fff-818c-e34321037686/volumes" Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.497117 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b409627-4716-4bdd-a660-4f88693150a2" path="/var/lib/kubelet/pods/4b409627-4716-4bdd-a660-4f88693150a2/volumes" Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.669019 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-796947dbf8-npktj"] Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.967938 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-796947dbf8-npktj" event={"ID":"cf16c324-2ece-43f9-8822-2e7ce9f06e7a","Type":"ContainerStarted","Data":"4ad7b656358d24f7acfe504e66e7ae4564ba521bdf7de20ef3541073974454a3"} Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.971169 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-767f84-ktpsj" event={"ID":"3a19d1ed-a1c3-4060-a254-a29eefa188bc","Type":"ContainerStarted","Data":"bd4017544baf3369c2a90bbc5bd5b4c1a71f4f181222634de612985d067e95b3"} Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.971413 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-767f84-ktpsj" Dec 06 15:51:32 crc kubenswrapper[4813]: I1206 15:51:32.982653 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-767f84-ktpsj" Dec 06 15:51:33 crc kubenswrapper[4813]: I1206 15:51:33.020335 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-767f84-ktpsj" podStartSLOduration=3.020302998 podStartE2EDuration="3.020302998s" podCreationTimestamp="2025-12-06 15:51:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:51:32.994371171 +0000 UTC m=+332.885250797" watchObservedRunningTime="2025-12-06 15:51:33.020302998 +0000 UTC m=+332.911182614" Dec 06 15:51:33 crc kubenswrapper[4813]: I1206 15:51:33.038588 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xwq7l" Dec 06 15:51:33 crc kubenswrapper[4813]: I1206 15:51:33.232672 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ws47p" Dec 06 15:51:33 crc kubenswrapper[4813]: I1206 15:51:33.232740 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ws47p" Dec 06 15:51:33 crc kubenswrapper[4813]: I1206 15:51:33.289038 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ws47p" Dec 06 15:51:33 crc kubenswrapper[4813]: I1206 15:51:33.828741 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-gxlzc"] Dec 06 15:51:33 crc kubenswrapper[4813]: I1206 15:51:33.829646 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-gxlzc" Dec 06 15:51:33 crc kubenswrapper[4813]: I1206 15:51:33.848837 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-gxlzc"] Dec 06 15:51:33 crc kubenswrapper[4813]: I1206 15:51:33.973830 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9b089579-93c8-47f2-909c-863d34d1c64c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-gxlzc\" (UID: \"9b089579-93c8-47f2-909c-863d34d1c64c\") " pod="openshift-image-registry/image-registry-66df7c8f76-gxlzc" Dec 06 15:51:33 crc kubenswrapper[4813]: I1206 15:51:33.973873 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84vgc\" (UniqueName: \"kubernetes.io/projected/9b089579-93c8-47f2-909c-863d34d1c64c-kube-api-access-84vgc\") pod \"image-registry-66df7c8f76-gxlzc\" (UID: \"9b089579-93c8-47f2-909c-863d34d1c64c\") " pod="openshift-image-registry/image-registry-66df7c8f76-gxlzc" Dec 06 15:51:33 crc kubenswrapper[4813]: I1206 15:51:33.973895 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9b089579-93c8-47f2-909c-863d34d1c64c-registry-tls\") pod \"image-registry-66df7c8f76-gxlzc\" (UID: \"9b089579-93c8-47f2-909c-863d34d1c64c\") " pod="openshift-image-registry/image-registry-66df7c8f76-gxlzc" Dec 06 15:51:33 crc kubenswrapper[4813]: I1206 15:51:33.973911 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9b089579-93c8-47f2-909c-863d34d1c64c-bound-sa-token\") pod \"image-registry-66df7c8f76-gxlzc\" (UID: \"9b089579-93c8-47f2-909c-863d34d1c64c\") " pod="openshift-image-registry/image-registry-66df7c8f76-gxlzc" Dec 06 15:51:33 crc kubenswrapper[4813]: I1206 15:51:33.973930 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9b089579-93c8-47f2-909c-863d34d1c64c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-gxlzc\" (UID: \"9b089579-93c8-47f2-909c-863d34d1c64c\") " pod="openshift-image-registry/image-registry-66df7c8f76-gxlzc" Dec 06 15:51:33 crc kubenswrapper[4813]: I1206 15:51:33.973948 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9b089579-93c8-47f2-909c-863d34d1c64c-registry-certificates\") pod \"image-registry-66df7c8f76-gxlzc\" (UID: \"9b089579-93c8-47f2-909c-863d34d1c64c\") " pod="openshift-image-registry/image-registry-66df7c8f76-gxlzc" Dec 06 15:51:33 crc kubenswrapper[4813]: I1206 15:51:33.974237 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-gxlzc\" (UID: \"9b089579-93c8-47f2-909c-863d34d1c64c\") " pod="openshift-image-registry/image-registry-66df7c8f76-gxlzc" Dec 06 15:51:33 crc kubenswrapper[4813]: I1206 15:51:33.974322 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9b089579-93c8-47f2-909c-863d34d1c64c-trusted-ca\") pod \"image-registry-66df7c8f76-gxlzc\" (UID: \"9b089579-93c8-47f2-909c-863d34d1c64c\") " pod="openshift-image-registry/image-registry-66df7c8f76-gxlzc" Dec 06 15:51:33 crc kubenswrapper[4813]: I1206 15:51:33.976672 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-796947dbf8-npktj" event={"ID":"cf16c324-2ece-43f9-8822-2e7ce9f06e7a","Type":"ContainerStarted","Data":"51326ae4e47533c90798d6c19892c4747153d67e3777e4e35de630a8c0b4f9b3"} Dec 06 15:51:33 crc kubenswrapper[4813]: I1206 15:51:33.977733 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-796947dbf8-npktj" Dec 06 15:51:33 crc kubenswrapper[4813]: I1206 15:51:33.995864 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-796947dbf8-npktj" Dec 06 15:51:33 crc kubenswrapper[4813]: I1206 15:51:33.999005 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-796947dbf8-npktj" podStartSLOduration=3.998988179 podStartE2EDuration="3.998988179s" podCreationTimestamp="2025-12-06 15:51:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:51:33.997604941 +0000 UTC m=+333.888484517" watchObservedRunningTime="2025-12-06 15:51:33.998988179 +0000 UTC m=+333.889867755" Dec 06 15:51:34 crc kubenswrapper[4813]: I1206 15:51:34.048493 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ws47p" Dec 06 15:51:34 crc kubenswrapper[4813]: I1206 15:51:34.071400 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-gxlzc\" (UID: \"9b089579-93c8-47f2-909c-863d34d1c64c\") " pod="openshift-image-registry/image-registry-66df7c8f76-gxlzc" Dec 06 15:51:34 crc kubenswrapper[4813]: I1206 15:51:34.075440 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9b089579-93c8-47f2-909c-863d34d1c64c-trusted-ca\") pod \"image-registry-66df7c8f76-gxlzc\" (UID: \"9b089579-93c8-47f2-909c-863d34d1c64c\") " pod="openshift-image-registry/image-registry-66df7c8f76-gxlzc" Dec 06 15:51:34 crc kubenswrapper[4813]: I1206 15:51:34.075526 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9b089579-93c8-47f2-909c-863d34d1c64c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-gxlzc\" (UID: \"9b089579-93c8-47f2-909c-863d34d1c64c\") " pod="openshift-image-registry/image-registry-66df7c8f76-gxlzc" Dec 06 15:51:34 crc kubenswrapper[4813]: I1206 15:51:34.075559 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84vgc\" (UniqueName: \"kubernetes.io/projected/9b089579-93c8-47f2-909c-863d34d1c64c-kube-api-access-84vgc\") pod \"image-registry-66df7c8f76-gxlzc\" (UID: \"9b089579-93c8-47f2-909c-863d34d1c64c\") " pod="openshift-image-registry/image-registry-66df7c8f76-gxlzc" Dec 06 15:51:34 crc kubenswrapper[4813]: I1206 15:51:34.075576 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9b089579-93c8-47f2-909c-863d34d1c64c-registry-tls\") pod \"image-registry-66df7c8f76-gxlzc\" (UID: \"9b089579-93c8-47f2-909c-863d34d1c64c\") " pod="openshift-image-registry/image-registry-66df7c8f76-gxlzc" Dec 06 15:51:34 crc kubenswrapper[4813]: I1206 15:51:34.075598 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9b089579-93c8-47f2-909c-863d34d1c64c-bound-sa-token\") pod \"image-registry-66df7c8f76-gxlzc\" (UID: \"9b089579-93c8-47f2-909c-863d34d1c64c\") " pod="openshift-image-registry/image-registry-66df7c8f76-gxlzc" Dec 06 15:51:34 crc kubenswrapper[4813]: I1206 15:51:34.075616 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9b089579-93c8-47f2-909c-863d34d1c64c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-gxlzc\" (UID: \"9b089579-93c8-47f2-909c-863d34d1c64c\") " pod="openshift-image-registry/image-registry-66df7c8f76-gxlzc" Dec 06 15:51:34 crc kubenswrapper[4813]: I1206 15:51:34.075631 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9b089579-93c8-47f2-909c-863d34d1c64c-registry-certificates\") pod \"image-registry-66df7c8f76-gxlzc\" (UID: \"9b089579-93c8-47f2-909c-863d34d1c64c\") " pod="openshift-image-registry/image-registry-66df7c8f76-gxlzc" Dec 06 15:51:34 crc kubenswrapper[4813]: I1206 15:51:34.079796 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9b089579-93c8-47f2-909c-863d34d1c64c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-gxlzc\" (UID: \"9b089579-93c8-47f2-909c-863d34d1c64c\") " pod="openshift-image-registry/image-registry-66df7c8f76-gxlzc" Dec 06 15:51:34 crc kubenswrapper[4813]: I1206 15:51:34.084290 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9b089579-93c8-47f2-909c-863d34d1c64c-trusted-ca\") pod \"image-registry-66df7c8f76-gxlzc\" (UID: \"9b089579-93c8-47f2-909c-863d34d1c64c\") " pod="openshift-image-registry/image-registry-66df7c8f76-gxlzc" Dec 06 15:51:34 crc kubenswrapper[4813]: I1206 15:51:34.084824 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9b089579-93c8-47f2-909c-863d34d1c64c-registry-certificates\") pod \"image-registry-66df7c8f76-gxlzc\" (UID: \"9b089579-93c8-47f2-909c-863d34d1c64c\") " pod="openshift-image-registry/image-registry-66df7c8f76-gxlzc" Dec 06 15:51:34 crc kubenswrapper[4813]: I1206 15:51:34.093055 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9b089579-93c8-47f2-909c-863d34d1c64c-registry-tls\") pod \"image-registry-66df7c8f76-gxlzc\" (UID: \"9b089579-93c8-47f2-909c-863d34d1c64c\") " pod="openshift-image-registry/image-registry-66df7c8f76-gxlzc" Dec 06 15:51:34 crc kubenswrapper[4813]: I1206 15:51:34.101494 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9b089579-93c8-47f2-909c-863d34d1c64c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-gxlzc\" (UID: \"9b089579-93c8-47f2-909c-863d34d1c64c\") " pod="openshift-image-registry/image-registry-66df7c8f76-gxlzc" Dec 06 15:51:34 crc kubenswrapper[4813]: I1206 15:51:34.104447 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9b089579-93c8-47f2-909c-863d34d1c64c-bound-sa-token\") pod \"image-registry-66df7c8f76-gxlzc\" (UID: \"9b089579-93c8-47f2-909c-863d34d1c64c\") " pod="openshift-image-registry/image-registry-66df7c8f76-gxlzc" Dec 06 15:51:34 crc kubenswrapper[4813]: I1206 15:51:34.113605 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84vgc\" (UniqueName: \"kubernetes.io/projected/9b089579-93c8-47f2-909c-863d34d1c64c-kube-api-access-84vgc\") pod \"image-registry-66df7c8f76-gxlzc\" (UID: \"9b089579-93c8-47f2-909c-863d34d1c64c\") " pod="openshift-image-registry/image-registry-66df7c8f76-gxlzc" Dec 06 15:51:34 crc kubenswrapper[4813]: I1206 15:51:34.142870 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-gxlzc" Dec 06 15:51:34 crc kubenswrapper[4813]: I1206 15:51:34.545879 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-gxlzc"] Dec 06 15:51:34 crc kubenswrapper[4813]: W1206 15:51:34.551590 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b089579_93c8_47f2_909c_863d34d1c64c.slice/crio-83862d8b3b2beee36251e677bc6bb07de5111e803b7c2ab7b5760bfd220261db WatchSource:0}: Error finding container 83862d8b3b2beee36251e677bc6bb07de5111e803b7c2ab7b5760bfd220261db: Status 404 returned error can't find the container with id 83862d8b3b2beee36251e677bc6bb07de5111e803b7c2ab7b5760bfd220261db Dec 06 15:51:34 crc kubenswrapper[4813]: I1206 15:51:34.894554 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-crqpg" Dec 06 15:51:34 crc kubenswrapper[4813]: I1206 15:51:34.894824 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-crqpg" Dec 06 15:51:34 crc kubenswrapper[4813]: I1206 15:51:34.952015 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-crqpg" Dec 06 15:51:34 crc kubenswrapper[4813]: I1206 15:51:34.983328 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-gxlzc" event={"ID":"9b089579-93c8-47f2-909c-863d34d1c64c","Type":"ContainerStarted","Data":"426b60675fb834ca6ba00ee9dedf2de1cd3405571c57e32aa1014364daf9627c"} Dec 06 15:51:34 crc kubenswrapper[4813]: I1206 15:51:34.984364 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-gxlzc" event={"ID":"9b089579-93c8-47f2-909c-863d34d1c64c","Type":"ContainerStarted","Data":"83862d8b3b2beee36251e677bc6bb07de5111e803b7c2ab7b5760bfd220261db"} Dec 06 15:51:35 crc kubenswrapper[4813]: I1206 15:51:35.020629 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-gxlzc" podStartSLOduration=2.020607641 podStartE2EDuration="2.020607641s" podCreationTimestamp="2025-12-06 15:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:51:35.014558716 +0000 UTC m=+334.905438302" watchObservedRunningTime="2025-12-06 15:51:35.020607641 +0000 UTC m=+334.911487217" Dec 06 15:51:35 crc kubenswrapper[4813]: I1206 15:51:35.039202 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-crqpg" Dec 06 15:51:35 crc kubenswrapper[4813]: I1206 15:51:35.081020 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-28zj9" Dec 06 15:51:35 crc kubenswrapper[4813]: I1206 15:51:35.095789 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-28zj9" Dec 06 15:51:35 crc kubenswrapper[4813]: I1206 15:51:35.138798 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-28zj9" Dec 06 15:51:35 crc kubenswrapper[4813]: I1206 15:51:35.990686 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-gxlzc" Dec 06 15:51:36 crc kubenswrapper[4813]: I1206 15:51:36.045445 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-28zj9" Dec 06 15:51:54 crc kubenswrapper[4813]: I1206 15:51:54.152741 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-gxlzc" Dec 06 15:51:54 crc kubenswrapper[4813]: I1206 15:51:54.255955 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-j79t8"] Dec 06 15:52:14 crc kubenswrapper[4813]: I1206 15:52:14.476393 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-796947dbf8-npktj"] Dec 06 15:52:14 crc kubenswrapper[4813]: I1206 15:52:14.477162 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-796947dbf8-npktj" podUID="cf16c324-2ece-43f9-8822-2e7ce9f06e7a" containerName="controller-manager" containerID="cri-o://51326ae4e47533c90798d6c19892c4747153d67e3777e4e35de630a8c0b4f9b3" gracePeriod=30 Dec 06 15:52:14 crc kubenswrapper[4813]: E1206 15:52:14.674657 4813 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcf16c324_2ece_43f9_8822_2e7ce9f06e7a.slice/crio-conmon-51326ae4e47533c90798d6c19892c4747153d67e3777e4e35de630a8c0b4f9b3.scope\": RecentStats: unable to find data in memory cache]" Dec 06 15:52:14 crc kubenswrapper[4813]: I1206 15:52:14.853834 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-796947dbf8-npktj" Dec 06 15:52:15 crc kubenswrapper[4813]: I1206 15:52:15.045594 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgtx4\" (UniqueName: \"kubernetes.io/projected/cf16c324-2ece-43f9-8822-2e7ce9f06e7a-kube-api-access-dgtx4\") pod \"cf16c324-2ece-43f9-8822-2e7ce9f06e7a\" (UID: \"cf16c324-2ece-43f9-8822-2e7ce9f06e7a\") " Dec 06 15:52:15 crc kubenswrapper[4813]: I1206 15:52:15.046346 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf16c324-2ece-43f9-8822-2e7ce9f06e7a-config\") pod \"cf16c324-2ece-43f9-8822-2e7ce9f06e7a\" (UID: \"cf16c324-2ece-43f9-8822-2e7ce9f06e7a\") " Dec 06 15:52:15 crc kubenswrapper[4813]: I1206 15:52:15.046547 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cf16c324-2ece-43f9-8822-2e7ce9f06e7a-client-ca\") pod \"cf16c324-2ece-43f9-8822-2e7ce9f06e7a\" (UID: \"cf16c324-2ece-43f9-8822-2e7ce9f06e7a\") " Dec 06 15:52:15 crc kubenswrapper[4813]: I1206 15:52:15.046747 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cf16c324-2ece-43f9-8822-2e7ce9f06e7a-proxy-ca-bundles\") pod \"cf16c324-2ece-43f9-8822-2e7ce9f06e7a\" (UID: \"cf16c324-2ece-43f9-8822-2e7ce9f06e7a\") " Dec 06 15:52:15 crc kubenswrapper[4813]: I1206 15:52:15.047522 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf16c324-2ece-43f9-8822-2e7ce9f06e7a-serving-cert\") pod \"cf16c324-2ece-43f9-8822-2e7ce9f06e7a\" (UID: \"cf16c324-2ece-43f9-8822-2e7ce9f06e7a\") " Dec 06 15:52:15 crc kubenswrapper[4813]: I1206 15:52:15.047398 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf16c324-2ece-43f9-8822-2e7ce9f06e7a-config" (OuterVolumeSpecName: "config") pod "cf16c324-2ece-43f9-8822-2e7ce9f06e7a" (UID: "cf16c324-2ece-43f9-8822-2e7ce9f06e7a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:52:15 crc kubenswrapper[4813]: I1206 15:52:15.047458 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf16c324-2ece-43f9-8822-2e7ce9f06e7a-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "cf16c324-2ece-43f9-8822-2e7ce9f06e7a" (UID: "cf16c324-2ece-43f9-8822-2e7ce9f06e7a"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:52:15 crc kubenswrapper[4813]: I1206 15:52:15.047448 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf16c324-2ece-43f9-8822-2e7ce9f06e7a-client-ca" (OuterVolumeSpecName: "client-ca") pod "cf16c324-2ece-43f9-8822-2e7ce9f06e7a" (UID: "cf16c324-2ece-43f9-8822-2e7ce9f06e7a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:52:15 crc kubenswrapper[4813]: I1206 15:52:15.048405 4813 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cf16c324-2ece-43f9-8822-2e7ce9f06e7a-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 15:52:15 crc kubenswrapper[4813]: I1206 15:52:15.048574 4813 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cf16c324-2ece-43f9-8822-2e7ce9f06e7a-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 06 15:52:15 crc kubenswrapper[4813]: I1206 15:52:15.048666 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf16c324-2ece-43f9-8822-2e7ce9f06e7a-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:52:15 crc kubenswrapper[4813]: I1206 15:52:15.057456 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf16c324-2ece-43f9-8822-2e7ce9f06e7a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "cf16c324-2ece-43f9-8822-2e7ce9f06e7a" (UID: "cf16c324-2ece-43f9-8822-2e7ce9f06e7a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:52:15 crc kubenswrapper[4813]: I1206 15:52:15.057530 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf16c324-2ece-43f9-8822-2e7ce9f06e7a-kube-api-access-dgtx4" (OuterVolumeSpecName: "kube-api-access-dgtx4") pod "cf16c324-2ece-43f9-8822-2e7ce9f06e7a" (UID: "cf16c324-2ece-43f9-8822-2e7ce9f06e7a"). InnerVolumeSpecName "kube-api-access-dgtx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:52:15 crc kubenswrapper[4813]: I1206 15:52:15.150055 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf16c324-2ece-43f9-8822-2e7ce9f06e7a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:52:15 crc kubenswrapper[4813]: I1206 15:52:15.150345 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgtx4\" (UniqueName: \"kubernetes.io/projected/cf16c324-2ece-43f9-8822-2e7ce9f06e7a-kube-api-access-dgtx4\") on node \"crc\" DevicePath \"\"" Dec 06 15:52:15 crc kubenswrapper[4813]: I1206 15:52:15.249146 4813 generic.go:334] "Generic (PLEG): container finished" podID="cf16c324-2ece-43f9-8822-2e7ce9f06e7a" containerID="51326ae4e47533c90798d6c19892c4747153d67e3777e4e35de630a8c0b4f9b3" exitCode=0 Dec 06 15:52:15 crc kubenswrapper[4813]: I1206 15:52:15.249308 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-796947dbf8-npktj" event={"ID":"cf16c324-2ece-43f9-8822-2e7ce9f06e7a","Type":"ContainerDied","Data":"51326ae4e47533c90798d6c19892c4747153d67e3777e4e35de630a8c0b4f9b3"} Dec 06 15:52:15 crc kubenswrapper[4813]: I1206 15:52:15.249355 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-796947dbf8-npktj" event={"ID":"cf16c324-2ece-43f9-8822-2e7ce9f06e7a","Type":"ContainerDied","Data":"4ad7b656358d24f7acfe504e66e7ae4564ba521bdf7de20ef3541073974454a3"} Dec 06 15:52:15 crc kubenswrapper[4813]: I1206 15:52:15.249386 4813 scope.go:117] "RemoveContainer" containerID="51326ae4e47533c90798d6c19892c4747153d67e3777e4e35de630a8c0b4f9b3" Dec 06 15:52:15 crc kubenswrapper[4813]: I1206 15:52:15.249609 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-796947dbf8-npktj" Dec 06 15:52:15 crc kubenswrapper[4813]: I1206 15:52:15.274366 4813 scope.go:117] "RemoveContainer" containerID="51326ae4e47533c90798d6c19892c4747153d67e3777e4e35de630a8c0b4f9b3" Dec 06 15:52:15 crc kubenswrapper[4813]: E1206 15:52:15.274877 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51326ae4e47533c90798d6c19892c4747153d67e3777e4e35de630a8c0b4f9b3\": container with ID starting with 51326ae4e47533c90798d6c19892c4747153d67e3777e4e35de630a8c0b4f9b3 not found: ID does not exist" containerID="51326ae4e47533c90798d6c19892c4747153d67e3777e4e35de630a8c0b4f9b3" Dec 06 15:52:15 crc kubenswrapper[4813]: I1206 15:52:15.274955 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51326ae4e47533c90798d6c19892c4747153d67e3777e4e35de630a8c0b4f9b3"} err="failed to get container status \"51326ae4e47533c90798d6c19892c4747153d67e3777e4e35de630a8c0b4f9b3\": rpc error: code = NotFound desc = could not find container \"51326ae4e47533c90798d6c19892c4747153d67e3777e4e35de630a8c0b4f9b3\": container with ID starting with 51326ae4e47533c90798d6c19892c4747153d67e3777e4e35de630a8c0b4f9b3 not found: ID does not exist" Dec 06 15:52:15 crc kubenswrapper[4813]: I1206 15:52:15.293104 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-796947dbf8-npktj"] Dec 06 15:52:15 crc kubenswrapper[4813]: I1206 15:52:15.297962 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-796947dbf8-npktj"] Dec 06 15:52:16 crc kubenswrapper[4813]: I1206 15:52:16.118240 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5f7b959576-rcmq4"] Dec 06 15:52:16 crc kubenswrapper[4813]: E1206 15:52:16.118655 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf16c324-2ece-43f9-8822-2e7ce9f06e7a" containerName="controller-manager" Dec 06 15:52:16 crc kubenswrapper[4813]: I1206 15:52:16.118681 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf16c324-2ece-43f9-8822-2e7ce9f06e7a" containerName="controller-manager" Dec 06 15:52:16 crc kubenswrapper[4813]: I1206 15:52:16.118856 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf16c324-2ece-43f9-8822-2e7ce9f06e7a" containerName="controller-manager" Dec 06 15:52:16 crc kubenswrapper[4813]: I1206 15:52:16.119520 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f7b959576-rcmq4" Dec 06 15:52:16 crc kubenswrapper[4813]: I1206 15:52:16.123725 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 06 15:52:16 crc kubenswrapper[4813]: I1206 15:52:16.124103 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 06 15:52:16 crc kubenswrapper[4813]: I1206 15:52:16.126596 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 06 15:52:16 crc kubenswrapper[4813]: I1206 15:52:16.126960 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 06 15:52:16 crc kubenswrapper[4813]: I1206 15:52:16.127754 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 06 15:52:16 crc kubenswrapper[4813]: I1206 15:52:16.129860 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 06 15:52:16 crc kubenswrapper[4813]: I1206 15:52:16.144407 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 06 15:52:16 crc kubenswrapper[4813]: I1206 15:52:16.153913 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5f7b959576-rcmq4"] Dec 06 15:52:16 crc kubenswrapper[4813]: I1206 15:52:16.166770 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d954196d-87bb-466b-b805-65201cfc42af-proxy-ca-bundles\") pod \"controller-manager-5f7b959576-rcmq4\" (UID: \"d954196d-87bb-466b-b805-65201cfc42af\") " pod="openshift-controller-manager/controller-manager-5f7b959576-rcmq4" Dec 06 15:52:16 crc kubenswrapper[4813]: I1206 15:52:16.166933 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjrbv\" (UniqueName: \"kubernetes.io/projected/d954196d-87bb-466b-b805-65201cfc42af-kube-api-access-zjrbv\") pod \"controller-manager-5f7b959576-rcmq4\" (UID: \"d954196d-87bb-466b-b805-65201cfc42af\") " pod="openshift-controller-manager/controller-manager-5f7b959576-rcmq4" Dec 06 15:52:16 crc kubenswrapper[4813]: I1206 15:52:16.167093 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d954196d-87bb-466b-b805-65201cfc42af-serving-cert\") pod \"controller-manager-5f7b959576-rcmq4\" (UID: \"d954196d-87bb-466b-b805-65201cfc42af\") " pod="openshift-controller-manager/controller-manager-5f7b959576-rcmq4" Dec 06 15:52:16 crc kubenswrapper[4813]: I1206 15:52:16.167178 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d954196d-87bb-466b-b805-65201cfc42af-client-ca\") pod \"controller-manager-5f7b959576-rcmq4\" (UID: \"d954196d-87bb-466b-b805-65201cfc42af\") " pod="openshift-controller-manager/controller-manager-5f7b959576-rcmq4" Dec 06 15:52:16 crc kubenswrapper[4813]: I1206 15:52:16.167249 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d954196d-87bb-466b-b805-65201cfc42af-config\") pod \"controller-manager-5f7b959576-rcmq4\" (UID: \"d954196d-87bb-466b-b805-65201cfc42af\") " pod="openshift-controller-manager/controller-manager-5f7b959576-rcmq4" Dec 06 15:52:16 crc kubenswrapper[4813]: I1206 15:52:16.269058 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d954196d-87bb-466b-b805-65201cfc42af-serving-cert\") pod \"controller-manager-5f7b959576-rcmq4\" (UID: \"d954196d-87bb-466b-b805-65201cfc42af\") " pod="openshift-controller-manager/controller-manager-5f7b959576-rcmq4" Dec 06 15:52:16 crc kubenswrapper[4813]: I1206 15:52:16.269137 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d954196d-87bb-466b-b805-65201cfc42af-client-ca\") pod \"controller-manager-5f7b959576-rcmq4\" (UID: \"d954196d-87bb-466b-b805-65201cfc42af\") " pod="openshift-controller-manager/controller-manager-5f7b959576-rcmq4" Dec 06 15:52:16 crc kubenswrapper[4813]: I1206 15:52:16.269176 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d954196d-87bb-466b-b805-65201cfc42af-config\") pod \"controller-manager-5f7b959576-rcmq4\" (UID: \"d954196d-87bb-466b-b805-65201cfc42af\") " pod="openshift-controller-manager/controller-manager-5f7b959576-rcmq4" Dec 06 15:52:16 crc kubenswrapper[4813]: I1206 15:52:16.269219 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d954196d-87bb-466b-b805-65201cfc42af-proxy-ca-bundles\") pod \"controller-manager-5f7b959576-rcmq4\" (UID: \"d954196d-87bb-466b-b805-65201cfc42af\") " pod="openshift-controller-manager/controller-manager-5f7b959576-rcmq4" Dec 06 15:52:16 crc kubenswrapper[4813]: I1206 15:52:16.269291 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjrbv\" (UniqueName: \"kubernetes.io/projected/d954196d-87bb-466b-b805-65201cfc42af-kube-api-access-zjrbv\") pod \"controller-manager-5f7b959576-rcmq4\" (UID: \"d954196d-87bb-466b-b805-65201cfc42af\") " pod="openshift-controller-manager/controller-manager-5f7b959576-rcmq4" Dec 06 15:52:16 crc kubenswrapper[4813]: I1206 15:52:16.270858 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d954196d-87bb-466b-b805-65201cfc42af-client-ca\") pod \"controller-manager-5f7b959576-rcmq4\" (UID: \"d954196d-87bb-466b-b805-65201cfc42af\") " pod="openshift-controller-manager/controller-manager-5f7b959576-rcmq4" Dec 06 15:52:16 crc kubenswrapper[4813]: I1206 15:52:16.271020 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d954196d-87bb-466b-b805-65201cfc42af-config\") pod \"controller-manager-5f7b959576-rcmq4\" (UID: \"d954196d-87bb-466b-b805-65201cfc42af\") " pod="openshift-controller-manager/controller-manager-5f7b959576-rcmq4" Dec 06 15:52:16 crc kubenswrapper[4813]: I1206 15:52:16.272167 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d954196d-87bb-466b-b805-65201cfc42af-proxy-ca-bundles\") pod \"controller-manager-5f7b959576-rcmq4\" (UID: \"d954196d-87bb-466b-b805-65201cfc42af\") " pod="openshift-controller-manager/controller-manager-5f7b959576-rcmq4" Dec 06 15:52:16 crc kubenswrapper[4813]: I1206 15:52:16.277107 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d954196d-87bb-466b-b805-65201cfc42af-serving-cert\") pod \"controller-manager-5f7b959576-rcmq4\" (UID: \"d954196d-87bb-466b-b805-65201cfc42af\") " pod="openshift-controller-manager/controller-manager-5f7b959576-rcmq4" Dec 06 15:52:16 crc kubenswrapper[4813]: I1206 15:52:16.298832 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjrbv\" (UniqueName: \"kubernetes.io/projected/d954196d-87bb-466b-b805-65201cfc42af-kube-api-access-zjrbv\") pod \"controller-manager-5f7b959576-rcmq4\" (UID: \"d954196d-87bb-466b-b805-65201cfc42af\") " pod="openshift-controller-manager/controller-manager-5f7b959576-rcmq4" Dec 06 15:52:16 crc kubenswrapper[4813]: I1206 15:52:16.447961 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f7b959576-rcmq4" Dec 06 15:52:16 crc kubenswrapper[4813]: I1206 15:52:16.503672 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf16c324-2ece-43f9-8822-2e7ce9f06e7a" path="/var/lib/kubelet/pods/cf16c324-2ece-43f9-8822-2e7ce9f06e7a/volumes" Dec 06 15:52:16 crc kubenswrapper[4813]: I1206 15:52:16.985237 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5f7b959576-rcmq4"] Dec 06 15:52:17 crc kubenswrapper[4813]: I1206 15:52:17.265972 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f7b959576-rcmq4" event={"ID":"d954196d-87bb-466b-b805-65201cfc42af","Type":"ContainerStarted","Data":"c6a116018372ea80ac9d39b65dc6566902cd582248d2b7d2d61d6a7bdca0d61f"} Dec 06 15:52:17 crc kubenswrapper[4813]: I1206 15:52:17.267463 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f7b959576-rcmq4" event={"ID":"d954196d-87bb-466b-b805-65201cfc42af","Type":"ContainerStarted","Data":"f3a1ac54d9066573c2c0292e0fdb7eb6e7ec7ad84bdc105b396d9808d1cd0db1"} Dec 06 15:52:17 crc kubenswrapper[4813]: I1206 15:52:17.268217 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5f7b959576-rcmq4" Dec 06 15:52:17 crc kubenswrapper[4813]: I1206 15:52:17.276404 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5f7b959576-rcmq4" Dec 06 15:52:17 crc kubenswrapper[4813]: I1206 15:52:17.317404 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5f7b959576-rcmq4" podStartSLOduration=3.31738565 podStartE2EDuration="3.31738565s" podCreationTimestamp="2025-12-06 15:52:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:52:17.298059133 +0000 UTC m=+377.188938709" watchObservedRunningTime="2025-12-06 15:52:17.31738565 +0000 UTC m=+377.208265236" Dec 06 15:52:19 crc kubenswrapper[4813]: I1206 15:52:19.312559 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" podUID="d2857e44-0b98-4f13-baca-3330474d8b44" containerName="registry" containerID="cri-o://776db7e851680747dc3871ee8baa6b50b31a045938289c74570f864d7d66dbf8" gracePeriod=30 Dec 06 15:52:19 crc kubenswrapper[4813]: I1206 15:52:19.428145 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 15:52:19 crc kubenswrapper[4813]: I1206 15:52:19.428218 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 15:52:19 crc kubenswrapper[4813]: I1206 15:52:19.754361 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:52:19 crc kubenswrapper[4813]: I1206 15:52:19.834947 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d2857e44-0b98-4f13-baca-3330474d8b44-registry-certificates\") pod \"d2857e44-0b98-4f13-baca-3330474d8b44\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " Dec 06 15:52:19 crc kubenswrapper[4813]: I1206 15:52:19.835022 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d2857e44-0b98-4f13-baca-3330474d8b44-registry-tls\") pod \"d2857e44-0b98-4f13-baca-3330474d8b44\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " Dec 06 15:52:19 crc kubenswrapper[4813]: I1206 15:52:19.835062 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67mng\" (UniqueName: \"kubernetes.io/projected/d2857e44-0b98-4f13-baca-3330474d8b44-kube-api-access-67mng\") pod \"d2857e44-0b98-4f13-baca-3330474d8b44\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " Dec 06 15:52:19 crc kubenswrapper[4813]: I1206 15:52:19.835118 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d2857e44-0b98-4f13-baca-3330474d8b44-ca-trust-extracted\") pod \"d2857e44-0b98-4f13-baca-3330474d8b44\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " Dec 06 15:52:19 crc kubenswrapper[4813]: I1206 15:52:19.835368 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"d2857e44-0b98-4f13-baca-3330474d8b44\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " Dec 06 15:52:19 crc kubenswrapper[4813]: I1206 15:52:19.835484 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d2857e44-0b98-4f13-baca-3330474d8b44-trusted-ca\") pod \"d2857e44-0b98-4f13-baca-3330474d8b44\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " Dec 06 15:52:19 crc kubenswrapper[4813]: I1206 15:52:19.835531 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d2857e44-0b98-4f13-baca-3330474d8b44-installation-pull-secrets\") pod \"d2857e44-0b98-4f13-baca-3330474d8b44\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " Dec 06 15:52:19 crc kubenswrapper[4813]: I1206 15:52:19.835580 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d2857e44-0b98-4f13-baca-3330474d8b44-bound-sa-token\") pod \"d2857e44-0b98-4f13-baca-3330474d8b44\" (UID: \"d2857e44-0b98-4f13-baca-3330474d8b44\") " Dec 06 15:52:19 crc kubenswrapper[4813]: I1206 15:52:19.836014 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2857e44-0b98-4f13-baca-3330474d8b44-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "d2857e44-0b98-4f13-baca-3330474d8b44" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:52:19 crc kubenswrapper[4813]: I1206 15:52:19.836749 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2857e44-0b98-4f13-baca-3330474d8b44-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "d2857e44-0b98-4f13-baca-3330474d8b44" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:52:19 crc kubenswrapper[4813]: I1206 15:52:19.841008 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2857e44-0b98-4f13-baca-3330474d8b44-kube-api-access-67mng" (OuterVolumeSpecName: "kube-api-access-67mng") pod "d2857e44-0b98-4f13-baca-3330474d8b44" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44"). InnerVolumeSpecName "kube-api-access-67mng". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:52:19 crc kubenswrapper[4813]: I1206 15:52:19.842351 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2857e44-0b98-4f13-baca-3330474d8b44-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "d2857e44-0b98-4f13-baca-3330474d8b44" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:52:19 crc kubenswrapper[4813]: I1206 15:52:19.842503 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2857e44-0b98-4f13-baca-3330474d8b44-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "d2857e44-0b98-4f13-baca-3330474d8b44" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:52:19 crc kubenswrapper[4813]: I1206 15:52:19.844413 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2857e44-0b98-4f13-baca-3330474d8b44-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "d2857e44-0b98-4f13-baca-3330474d8b44" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:52:19 crc kubenswrapper[4813]: I1206 15:52:19.852121 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "d2857e44-0b98-4f13-baca-3330474d8b44" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 06 15:52:19 crc kubenswrapper[4813]: I1206 15:52:19.860832 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2857e44-0b98-4f13-baca-3330474d8b44-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "d2857e44-0b98-4f13-baca-3330474d8b44" (UID: "d2857e44-0b98-4f13-baca-3330474d8b44"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:52:19 crc kubenswrapper[4813]: I1206 15:52:19.937058 4813 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d2857e44-0b98-4f13-baca-3330474d8b44-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 15:52:19 crc kubenswrapper[4813]: I1206 15:52:19.937087 4813 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d2857e44-0b98-4f13-baca-3330474d8b44-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 06 15:52:19 crc kubenswrapper[4813]: I1206 15:52:19.937098 4813 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d2857e44-0b98-4f13-baca-3330474d8b44-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 15:52:19 crc kubenswrapper[4813]: I1206 15:52:19.937106 4813 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d2857e44-0b98-4f13-baca-3330474d8b44-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 06 15:52:19 crc kubenswrapper[4813]: I1206 15:52:19.937115 4813 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d2857e44-0b98-4f13-baca-3330474d8b44-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 06 15:52:19 crc kubenswrapper[4813]: I1206 15:52:19.937124 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67mng\" (UniqueName: \"kubernetes.io/projected/d2857e44-0b98-4f13-baca-3330474d8b44-kube-api-access-67mng\") on node \"crc\" DevicePath \"\"" Dec 06 15:52:19 crc kubenswrapper[4813]: I1206 15:52:19.937133 4813 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d2857e44-0b98-4f13-baca-3330474d8b44-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 06 15:52:20 crc kubenswrapper[4813]: I1206 15:52:20.315333 4813 generic.go:334] "Generic (PLEG): container finished" podID="d2857e44-0b98-4f13-baca-3330474d8b44" containerID="776db7e851680747dc3871ee8baa6b50b31a045938289c74570f864d7d66dbf8" exitCode=0 Dec 06 15:52:20 crc kubenswrapper[4813]: I1206 15:52:20.315407 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" event={"ID":"d2857e44-0b98-4f13-baca-3330474d8b44","Type":"ContainerDied","Data":"776db7e851680747dc3871ee8baa6b50b31a045938289c74570f864d7d66dbf8"} Dec 06 15:52:20 crc kubenswrapper[4813]: I1206 15:52:20.315399 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" Dec 06 15:52:20 crc kubenswrapper[4813]: I1206 15:52:20.315454 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-j79t8" event={"ID":"d2857e44-0b98-4f13-baca-3330474d8b44","Type":"ContainerDied","Data":"4c61e905c9692c4d5c5e38a288853af45e48b19329159f50a06490cabdaf0f01"} Dec 06 15:52:20 crc kubenswrapper[4813]: I1206 15:52:20.315492 4813 scope.go:117] "RemoveContainer" containerID="776db7e851680747dc3871ee8baa6b50b31a045938289c74570f864d7d66dbf8" Dec 06 15:52:20 crc kubenswrapper[4813]: I1206 15:52:20.347940 4813 scope.go:117] "RemoveContainer" containerID="776db7e851680747dc3871ee8baa6b50b31a045938289c74570f864d7d66dbf8" Dec 06 15:52:20 crc kubenswrapper[4813]: E1206 15:52:20.348493 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"776db7e851680747dc3871ee8baa6b50b31a045938289c74570f864d7d66dbf8\": container with ID starting with 776db7e851680747dc3871ee8baa6b50b31a045938289c74570f864d7d66dbf8 not found: ID does not exist" containerID="776db7e851680747dc3871ee8baa6b50b31a045938289c74570f864d7d66dbf8" Dec 06 15:52:20 crc kubenswrapper[4813]: I1206 15:52:20.348546 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"776db7e851680747dc3871ee8baa6b50b31a045938289c74570f864d7d66dbf8"} err="failed to get container status \"776db7e851680747dc3871ee8baa6b50b31a045938289c74570f864d7d66dbf8\": rpc error: code = NotFound desc = could not find container \"776db7e851680747dc3871ee8baa6b50b31a045938289c74570f864d7d66dbf8\": container with ID starting with 776db7e851680747dc3871ee8baa6b50b31a045938289c74570f864d7d66dbf8 not found: ID does not exist" Dec 06 15:52:20 crc kubenswrapper[4813]: I1206 15:52:20.384048 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-j79t8"] Dec 06 15:52:20 crc kubenswrapper[4813]: I1206 15:52:20.388834 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-j79t8"] Dec 06 15:52:20 crc kubenswrapper[4813]: I1206 15:52:20.499432 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2857e44-0b98-4f13-baca-3330474d8b44" path="/var/lib/kubelet/pods/d2857e44-0b98-4f13-baca-3330474d8b44/volumes" Dec 06 15:52:49 crc kubenswrapper[4813]: I1206 15:52:49.428667 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 15:52:49 crc kubenswrapper[4813]: I1206 15:52:49.429427 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 15:53:19 crc kubenswrapper[4813]: I1206 15:53:19.427827 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 15:53:19 crc kubenswrapper[4813]: I1206 15:53:19.428431 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 15:53:19 crc kubenswrapper[4813]: I1206 15:53:19.428472 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" Dec 06 15:53:19 crc kubenswrapper[4813]: I1206 15:53:19.429044 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9bb3ab0a5b69e0227b05bdb28844955eb8df2d6874e1937ca3d3672f21c852e9"} pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 15:53:19 crc kubenswrapper[4813]: I1206 15:53:19.429105 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" containerID="cri-o://9bb3ab0a5b69e0227b05bdb28844955eb8df2d6874e1937ca3d3672f21c852e9" gracePeriod=600 Dec 06 15:53:19 crc kubenswrapper[4813]: I1206 15:53:19.713698 4813 generic.go:334] "Generic (PLEG): container finished" podID="d88e8bae-c055-4c55-b548-f621ff96de06" containerID="9bb3ab0a5b69e0227b05bdb28844955eb8df2d6874e1937ca3d3672f21c852e9" exitCode=0 Dec 06 15:53:19 crc kubenswrapper[4813]: I1206 15:53:19.713797 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerDied","Data":"9bb3ab0a5b69e0227b05bdb28844955eb8df2d6874e1937ca3d3672f21c852e9"} Dec 06 15:53:19 crc kubenswrapper[4813]: I1206 15:53:19.714000 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerStarted","Data":"6841d4ac1913266f58304ddc35c3d4b57d8eed36652e13d503003868ba785be0"} Dec 06 15:53:19 crc kubenswrapper[4813]: I1206 15:53:19.714021 4813 scope.go:117] "RemoveContainer" containerID="0d2626cc9bffe949668c6fb8631b1f7035078c99c4def0c422852193bac77624" Dec 06 15:55:19 crc kubenswrapper[4813]: I1206 15:55:19.428338 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 15:55:19 crc kubenswrapper[4813]: I1206 15:55:19.428763 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 15:55:49 crc kubenswrapper[4813]: I1206 15:55:49.428040 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 15:55:49 crc kubenswrapper[4813]: I1206 15:55:49.428838 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.164138 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-6jx97"] Dec 06 15:55:58 crc kubenswrapper[4813]: E1206 15:55:58.164900 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2857e44-0b98-4f13-baca-3330474d8b44" containerName="registry" Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.164915 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2857e44-0b98-4f13-baca-3330474d8b44" containerName="registry" Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.165014 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2857e44-0b98-4f13-baca-3330474d8b44" containerName="registry" Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.165472 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-6jx97" Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.170445 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.170660 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.172681 4813 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-q8dxc" Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.174872 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-gqwzr"] Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.175658 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-gqwzr" Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.178856 4813 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-29p7k" Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.186916 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2lbd\" (UniqueName: \"kubernetes.io/projected/93a0448f-03c2-4a73-8ebc-b6bc1a5a3cd3-kube-api-access-r2lbd\") pod \"cert-manager-cainjector-7f985d654d-6jx97\" (UID: \"93a0448f-03c2-4a73-8ebc-b6bc1a5a3cd3\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-6jx97" Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.186973 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twk9z\" (UniqueName: \"kubernetes.io/projected/9a87a677-1122-46a6-aae7-820387f122d5-kube-api-access-twk9z\") pod \"cert-manager-5b446d88c5-gqwzr\" (UID: \"9a87a677-1122-46a6-aae7-820387f122d5\") " pod="cert-manager/cert-manager-5b446d88c5-gqwzr" Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.190105 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-4m6ft"] Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.190885 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-4m6ft" Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.193854 4813 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-24wdd" Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.205303 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-6jx97"] Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.212739 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-4m6ft"] Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.238229 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-gqwzr"] Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.288453 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2lbd\" (UniqueName: \"kubernetes.io/projected/93a0448f-03c2-4a73-8ebc-b6bc1a5a3cd3-kube-api-access-r2lbd\") pod \"cert-manager-cainjector-7f985d654d-6jx97\" (UID: \"93a0448f-03c2-4a73-8ebc-b6bc1a5a3cd3\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-6jx97" Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.288526 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twk9z\" (UniqueName: \"kubernetes.io/projected/9a87a677-1122-46a6-aae7-820387f122d5-kube-api-access-twk9z\") pod \"cert-manager-5b446d88c5-gqwzr\" (UID: \"9a87a677-1122-46a6-aae7-820387f122d5\") " pod="cert-manager/cert-manager-5b446d88c5-gqwzr" Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.288609 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rk6jv\" (UniqueName: \"kubernetes.io/projected/b5fd38e1-0a32-486d-839d-e3cce8c150fe-kube-api-access-rk6jv\") pod \"cert-manager-webhook-5655c58dd6-4m6ft\" (UID: \"b5fd38e1-0a32-486d-839d-e3cce8c150fe\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-4m6ft" Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.308643 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twk9z\" (UniqueName: \"kubernetes.io/projected/9a87a677-1122-46a6-aae7-820387f122d5-kube-api-access-twk9z\") pod \"cert-manager-5b446d88c5-gqwzr\" (UID: \"9a87a677-1122-46a6-aae7-820387f122d5\") " pod="cert-manager/cert-manager-5b446d88c5-gqwzr" Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.315250 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2lbd\" (UniqueName: \"kubernetes.io/projected/93a0448f-03c2-4a73-8ebc-b6bc1a5a3cd3-kube-api-access-r2lbd\") pod \"cert-manager-cainjector-7f985d654d-6jx97\" (UID: \"93a0448f-03c2-4a73-8ebc-b6bc1a5a3cd3\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-6jx97" Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.389980 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rk6jv\" (UniqueName: \"kubernetes.io/projected/b5fd38e1-0a32-486d-839d-e3cce8c150fe-kube-api-access-rk6jv\") pod \"cert-manager-webhook-5655c58dd6-4m6ft\" (UID: \"b5fd38e1-0a32-486d-839d-e3cce8c150fe\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-4m6ft" Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.408240 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rk6jv\" (UniqueName: \"kubernetes.io/projected/b5fd38e1-0a32-486d-839d-e3cce8c150fe-kube-api-access-rk6jv\") pod \"cert-manager-webhook-5655c58dd6-4m6ft\" (UID: \"b5fd38e1-0a32-486d-839d-e3cce8c150fe\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-4m6ft" Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.484907 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-6jx97" Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.500597 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-gqwzr" Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.509684 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-4m6ft" Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.749567 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-gqwzr"] Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.760083 4813 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 15:55:58 crc kubenswrapper[4813]: I1206 15:55:58.822081 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-4m6ft"] Dec 06 15:55:59 crc kubenswrapper[4813]: I1206 15:55:59.059154 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-6jx97"] Dec 06 15:55:59 crc kubenswrapper[4813]: W1206 15:55:59.063553 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93a0448f_03c2_4a73_8ebc_b6bc1a5a3cd3.slice/crio-3312f28ec9f233fe7bf82ca86f12518f8b6f458499dd75a605511dda34938753 WatchSource:0}: Error finding container 3312f28ec9f233fe7bf82ca86f12518f8b6f458499dd75a605511dda34938753: Status 404 returned error can't find the container with id 3312f28ec9f233fe7bf82ca86f12518f8b6f458499dd75a605511dda34938753 Dec 06 15:55:59 crc kubenswrapper[4813]: I1206 15:55:59.138544 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-4m6ft" event={"ID":"b5fd38e1-0a32-486d-839d-e3cce8c150fe","Type":"ContainerStarted","Data":"c89af45b0bc4e99c0099f026d1c6ca767a5aa4c04a1749bbc2859985210117c3"} Dec 06 15:55:59 crc kubenswrapper[4813]: I1206 15:55:59.139193 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-gqwzr" event={"ID":"9a87a677-1122-46a6-aae7-820387f122d5","Type":"ContainerStarted","Data":"cc56ba72f5107a7dfe866fe9c81368b86f95b67aee54c14b4f790ad6cab4a8e2"} Dec 06 15:55:59 crc kubenswrapper[4813]: I1206 15:55:59.139809 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-6jx97" event={"ID":"93a0448f-03c2-4a73-8ebc-b6bc1a5a3cd3","Type":"ContainerStarted","Data":"3312f28ec9f233fe7bf82ca86f12518f8b6f458499dd75a605511dda34938753"} Dec 06 15:56:03 crc kubenswrapper[4813]: I1206 15:56:03.158314 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-4m6ft" event={"ID":"b5fd38e1-0a32-486d-839d-e3cce8c150fe","Type":"ContainerStarted","Data":"fc069a76b28c204f076d4248ac6dbd2bba7cad26603f2248154c51d63f024c27"} Dec 06 15:56:03 crc kubenswrapper[4813]: I1206 15:56:03.158515 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-4m6ft" Dec 06 15:56:03 crc kubenswrapper[4813]: I1206 15:56:03.159953 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-gqwzr" event={"ID":"9a87a677-1122-46a6-aae7-820387f122d5","Type":"ContainerStarted","Data":"f8cdb6cc12504dc80d75b995b09ade23a439b03fe843918c818c38d7a5dd8690"} Dec 06 15:56:03 crc kubenswrapper[4813]: I1206 15:56:03.162183 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-6jx97" event={"ID":"93a0448f-03c2-4a73-8ebc-b6bc1a5a3cd3","Type":"ContainerStarted","Data":"77bcdd537b6af18e16671bd098bdbd3a00bd1e6041b7dc40a2f986fe68936336"} Dec 06 15:56:03 crc kubenswrapper[4813]: I1206 15:56:03.185613 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-4m6ft" podStartSLOduration=1.9414131430000001 podStartE2EDuration="5.185594278s" podCreationTimestamp="2025-12-06 15:55:58 +0000 UTC" firstStartedPulling="2025-12-06 15:55:58.83116455 +0000 UTC m=+598.722044126" lastFinishedPulling="2025-12-06 15:56:02.075345685 +0000 UTC m=+601.966225261" observedRunningTime="2025-12-06 15:56:03.179633866 +0000 UTC m=+603.070513442" watchObservedRunningTime="2025-12-06 15:56:03.185594278 +0000 UTC m=+603.076473894" Dec 06 15:56:03 crc kubenswrapper[4813]: I1206 15:56:03.211228 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-6jx97" podStartSLOduration=2.14761572 podStartE2EDuration="5.211161132s" podCreationTimestamp="2025-12-06 15:55:58 +0000 UTC" firstStartedPulling="2025-12-06 15:55:59.065156342 +0000 UTC m=+598.956035918" lastFinishedPulling="2025-12-06 15:56:02.128701754 +0000 UTC m=+602.019581330" observedRunningTime="2025-12-06 15:56:03.204934083 +0000 UTC m=+603.095813699" watchObservedRunningTime="2025-12-06 15:56:03.211161132 +0000 UTC m=+603.102040738" Dec 06 15:56:03 crc kubenswrapper[4813]: I1206 15:56:03.225197 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-gqwzr" podStartSLOduration=1.909359542 podStartE2EDuration="5.225181682s" podCreationTimestamp="2025-12-06 15:55:58 +0000 UTC" firstStartedPulling="2025-12-06 15:55:58.759892705 +0000 UTC m=+598.650772281" lastFinishedPulling="2025-12-06 15:56:02.075714805 +0000 UTC m=+601.966594421" observedRunningTime="2025-12-06 15:56:03.223727683 +0000 UTC m=+603.114607299" watchObservedRunningTime="2025-12-06 15:56:03.225181682 +0000 UTC m=+603.116061288" Dec 06 15:56:08 crc kubenswrapper[4813]: I1206 15:56:08.512567 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-4m6ft" Dec 06 15:56:08 crc kubenswrapper[4813]: I1206 15:56:08.886059 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-xdt4d"] Dec 06 15:56:08 crc kubenswrapper[4813]: I1206 15:56:08.886781 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa" gracePeriod=30 Dec 06 15:56:08 crc kubenswrapper[4813]: I1206 15:56:08.886780 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="northd" containerID="cri-o://57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc" gracePeriod=30 Dec 06 15:56:08 crc kubenswrapper[4813]: I1206 15:56:08.886782 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="nbdb" containerID="cri-o://2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45" gracePeriod=30 Dec 06 15:56:08 crc kubenswrapper[4813]: I1206 15:56:08.886941 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="sbdb" containerID="cri-o://afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2" gracePeriod=30 Dec 06 15:56:08 crc kubenswrapper[4813]: I1206 15:56:08.887004 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="kube-rbac-proxy-node" containerID="cri-o://6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e" gracePeriod=30 Dec 06 15:56:08 crc kubenswrapper[4813]: I1206 15:56:08.887051 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="ovn-acl-logging" containerID="cri-o://f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c" gracePeriod=30 Dec 06 15:56:08 crc kubenswrapper[4813]: I1206 15:56:08.886723 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="ovn-controller" containerID="cri-o://5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c" gracePeriod=30 Dec 06 15:56:08 crc kubenswrapper[4813]: I1206 15:56:08.936019 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="ovnkube-controller" containerID="cri-o://5cf32df4a99c42b7b090412102796d4a6d265bc5a4de2ecf1648cadbac28ea73" gracePeriod=30 Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.204900 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gl495_856bb45b-525c-4ef0-bf7c-0691cf54b342/kube-multus/2.log" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.205326 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gl495_856bb45b-525c-4ef0-bf7c-0691cf54b342/kube-multus/1.log" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.205357 4813 generic.go:334] "Generic (PLEG): container finished" podID="856bb45b-525c-4ef0-bf7c-0691cf54b342" containerID="ef40f8e7e7629e97c914ef68c7d15da2e144cc9558102b30198efb0618912fe8" exitCode=2 Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.205400 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gl495" event={"ID":"856bb45b-525c-4ef0-bf7c-0691cf54b342","Type":"ContainerDied","Data":"ef40f8e7e7629e97c914ef68c7d15da2e144cc9558102b30198efb0618912fe8"} Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.205430 4813 scope.go:117] "RemoveContainer" containerID="7001bdf05c6acf52c9be35bd63ae103390bf04d2a00a41bf08d376dcdc497338" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.205849 4813 scope.go:117] "RemoveContainer" containerID="ef40f8e7e7629e97c914ef68c7d15da2e144cc9558102b30198efb0618912fe8" Dec 06 15:56:09 crc kubenswrapper[4813]: E1206 15:56:09.206013 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-gl495_openshift-multus(856bb45b-525c-4ef0-bf7c-0691cf54b342)\"" pod="openshift-multus/multus-gl495" podUID="856bb45b-525c-4ef0-bf7c-0691cf54b342" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.212620 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xdt4d_0e54e31f-e9f8-4dd5-8738-c785df002c93/ovnkube-controller/3.log" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.226104 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xdt4d_0e54e31f-e9f8-4dd5-8738-c785df002c93/ovn-acl-logging/0.log" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.226605 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xdt4d_0e54e31f-e9f8-4dd5-8738-c785df002c93/ovn-controller/0.log" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.227039 4813 generic.go:334] "Generic (PLEG): container finished" podID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerID="5cf32df4a99c42b7b090412102796d4a6d265bc5a4de2ecf1648cadbac28ea73" exitCode=0 Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.227066 4813 generic.go:334] "Generic (PLEG): container finished" podID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerID="2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa" exitCode=0 Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.227074 4813 generic.go:334] "Generic (PLEG): container finished" podID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerID="6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e" exitCode=0 Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.227081 4813 generic.go:334] "Generic (PLEG): container finished" podID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerID="f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c" exitCode=143 Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.227087 4813 generic.go:334] "Generic (PLEG): container finished" podID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerID="5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c" exitCode=143 Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.227106 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" event={"ID":"0e54e31f-e9f8-4dd5-8738-c785df002c93","Type":"ContainerDied","Data":"5cf32df4a99c42b7b090412102796d4a6d265bc5a4de2ecf1648cadbac28ea73"} Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.227131 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" event={"ID":"0e54e31f-e9f8-4dd5-8738-c785df002c93","Type":"ContainerDied","Data":"2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa"} Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.227140 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" event={"ID":"0e54e31f-e9f8-4dd5-8738-c785df002c93","Type":"ContainerDied","Data":"6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e"} Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.227149 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" event={"ID":"0e54e31f-e9f8-4dd5-8738-c785df002c93","Type":"ContainerDied","Data":"f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c"} Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.227157 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" event={"ID":"0e54e31f-e9f8-4dd5-8738-c785df002c93","Type":"ContainerDied","Data":"5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c"} Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.327819 4813 scope.go:117] "RemoveContainer" containerID="6b654f2820e7e70a18617e577d1dffa10f2e6e786d8e8a2483493d1e2cbf32e6" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.589046 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xdt4d_0e54e31f-e9f8-4dd5-8738-c785df002c93/ovn-acl-logging/0.log" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.589686 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xdt4d_0e54e31f-e9f8-4dd5-8738-c785df002c93/ovn-controller/0.log" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.590234 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.672317 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-lnmrs"] Dec 06 15:56:09 crc kubenswrapper[4813]: E1206 15:56:09.672764 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="northd" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.672798 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="northd" Dec 06 15:56:09 crc kubenswrapper[4813]: E1206 15:56:09.672821 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="nbdb" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.672834 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="nbdb" Dec 06 15:56:09 crc kubenswrapper[4813]: E1206 15:56:09.672854 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="ovnkube-controller" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.672866 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="ovnkube-controller" Dec 06 15:56:09 crc kubenswrapper[4813]: E1206 15:56:09.672883 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="ovnkube-controller" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.672896 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="ovnkube-controller" Dec 06 15:56:09 crc kubenswrapper[4813]: E1206 15:56:09.672911 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="ovn-controller" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.672923 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="ovn-controller" Dec 06 15:56:09 crc kubenswrapper[4813]: E1206 15:56:09.672941 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="kube-rbac-proxy-node" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.672953 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="kube-rbac-proxy-node" Dec 06 15:56:09 crc kubenswrapper[4813]: E1206 15:56:09.672971 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="sbdb" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.672983 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="sbdb" Dec 06 15:56:09 crc kubenswrapper[4813]: E1206 15:56:09.672997 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="kubecfg-setup" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.673009 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="kubecfg-setup" Dec 06 15:56:09 crc kubenswrapper[4813]: E1206 15:56:09.673026 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="kube-rbac-proxy-ovn-metrics" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.673037 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="kube-rbac-proxy-ovn-metrics" Dec 06 15:56:09 crc kubenswrapper[4813]: E1206 15:56:09.673053 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="ovnkube-controller" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.673065 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="ovnkube-controller" Dec 06 15:56:09 crc kubenswrapper[4813]: E1206 15:56:09.673080 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="ovn-acl-logging" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.673092 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="ovn-acl-logging" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.673254 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="nbdb" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.673304 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="ovn-acl-logging" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.673318 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="ovnkube-controller" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.673332 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="northd" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.673346 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="sbdb" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.673363 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="ovn-controller" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.673379 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="kube-rbac-proxy-ovn-metrics" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.673391 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="ovnkube-controller" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.673408 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="ovnkube-controller" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.673422 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="kube-rbac-proxy-node" Dec 06 15:56:09 crc kubenswrapper[4813]: E1206 15:56:09.673575 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="ovnkube-controller" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.673590 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="ovnkube-controller" Dec 06 15:56:09 crc kubenswrapper[4813]: E1206 15:56:09.673615 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="ovnkube-controller" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.673626 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="ovnkube-controller" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.673793 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="ovnkube-controller" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.673814 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerName="ovnkube-controller" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.676985 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.679198 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-var-lib-openvswitch\") pod \"0e54e31f-e9f8-4dd5-8738-c785df002c93\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.679315 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0e54e31f-e9f8-4dd5-8738-c785df002c93-ovnkube-script-lib\") pod \"0e54e31f-e9f8-4dd5-8738-c785df002c93\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.679320 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "0e54e31f-e9f8-4dd5-8738-c785df002c93" (UID: "0e54e31f-e9f8-4dd5-8738-c785df002c93"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.679382 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-run-ovn-kubernetes\") pod \"0e54e31f-e9f8-4dd5-8738-c785df002c93\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.679422 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-run-ovn\") pod \"0e54e31f-e9f8-4dd5-8738-c785df002c93\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.679476 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "0e54e31f-e9f8-4dd5-8738-c785df002c93" (UID: "0e54e31f-e9f8-4dd5-8738-c785df002c93"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.679611 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "0e54e31f-e9f8-4dd5-8738-c785df002c93" (UID: "0e54e31f-e9f8-4dd5-8738-c785df002c93"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.679691 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-run-systemd\") pod \"0e54e31f-e9f8-4dd5-8738-c785df002c93\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.680130 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e54e31f-e9f8-4dd5-8738-c785df002c93-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "0e54e31f-e9f8-4dd5-8738-c785df002c93" (UID: "0e54e31f-e9f8-4dd5-8738-c785df002c93"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.681041 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-run-openvswitch\") pod \"0e54e31f-e9f8-4dd5-8738-c785df002c93\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.681397 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0e54e31f-e9f8-4dd5-8738-c785df002c93-ovn-node-metrics-cert\") pod \"0e54e31f-e9f8-4dd5-8738-c785df002c93\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.681430 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-cni-bin\") pod \"0e54e31f-e9f8-4dd5-8738-c785df002c93\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.681472 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0e54e31f-e9f8-4dd5-8738-c785df002c93-ovnkube-config\") pod \"0e54e31f-e9f8-4dd5-8738-c785df002c93\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.681470 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "0e54e31f-e9f8-4dd5-8738-c785df002c93" (UID: "0e54e31f-e9f8-4dd5-8738-c785df002c93"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.681549 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "0e54e31f-e9f8-4dd5-8738-c785df002c93" (UID: "0e54e31f-e9f8-4dd5-8738-c785df002c93"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.681647 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-systemd-units\") pod \"0e54e31f-e9f8-4dd5-8738-c785df002c93\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.681753 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "0e54e31f-e9f8-4dd5-8738-c785df002c93" (UID: "0e54e31f-e9f8-4dd5-8738-c785df002c93"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.681773 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-var-lib-cni-networks-ovn-kubernetes\") pod \"0e54e31f-e9f8-4dd5-8738-c785df002c93\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.681811 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-log-socket\") pod \"0e54e31f-e9f8-4dd5-8738-c785df002c93\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.681849 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbvf9\" (UniqueName: \"kubernetes.io/projected/0e54e31f-e9f8-4dd5-8738-c785df002c93-kube-api-access-nbvf9\") pod \"0e54e31f-e9f8-4dd5-8738-c785df002c93\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.681850 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "0e54e31f-e9f8-4dd5-8738-c785df002c93" (UID: "0e54e31f-e9f8-4dd5-8738-c785df002c93"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.681879 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-run-netns\") pod \"0e54e31f-e9f8-4dd5-8738-c785df002c93\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.681897 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e54e31f-e9f8-4dd5-8738-c785df002c93-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "0e54e31f-e9f8-4dd5-8738-c785df002c93" (UID: "0e54e31f-e9f8-4dd5-8738-c785df002c93"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.681918 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-etc-openvswitch\") pod \"0e54e31f-e9f8-4dd5-8738-c785df002c93\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.681987 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-cni-netd\") pod \"0e54e31f-e9f8-4dd5-8738-c785df002c93\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.681918 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-log-socket" (OuterVolumeSpecName: "log-socket") pod "0e54e31f-e9f8-4dd5-8738-c785df002c93" (UID: "0e54e31f-e9f8-4dd5-8738-c785df002c93"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.682029 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-kubelet\") pod \"0e54e31f-e9f8-4dd5-8738-c785df002c93\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.682079 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-node-log\") pod \"0e54e31f-e9f8-4dd5-8738-c785df002c93\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.681940 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "0e54e31f-e9f8-4dd5-8738-c785df002c93" (UID: "0e54e31f-e9f8-4dd5-8738-c785df002c93"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.681960 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "0e54e31f-e9f8-4dd5-8738-c785df002c93" (UID: "0e54e31f-e9f8-4dd5-8738-c785df002c93"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.682069 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "0e54e31f-e9f8-4dd5-8738-c785df002c93" (UID: "0e54e31f-e9f8-4dd5-8738-c785df002c93"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.682123 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-slash\") pod \"0e54e31f-e9f8-4dd5-8738-c785df002c93\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.682115 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "0e54e31f-e9f8-4dd5-8738-c785df002c93" (UID: "0e54e31f-e9f8-4dd5-8738-c785df002c93"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.682156 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0e54e31f-e9f8-4dd5-8738-c785df002c93-env-overrides\") pod \"0e54e31f-e9f8-4dd5-8738-c785df002c93\" (UID: \"0e54e31f-e9f8-4dd5-8738-c785df002c93\") " Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.682160 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-node-log" (OuterVolumeSpecName: "node-log") pod "0e54e31f-e9f8-4dd5-8738-c785df002c93" (UID: "0e54e31f-e9f8-4dd5-8738-c785df002c93"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.682198 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-slash" (OuterVolumeSpecName: "host-slash") pod "0e54e31f-e9f8-4dd5-8738-c785df002c93" (UID: "0e54e31f-e9f8-4dd5-8738-c785df002c93"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.682576 4813 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.682602 4813 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.682625 4813 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-log-socket\") on node \"crc\" DevicePath \"\"" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.682647 4813 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.682669 4813 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.682687 4813 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.682805 4813 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.682850 4813 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-node-log\") on node \"crc\" DevicePath \"\"" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.682966 4813 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-slash\") on node \"crc\" DevicePath \"\"" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.682839 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e54e31f-e9f8-4dd5-8738-c785df002c93-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "0e54e31f-e9f8-4dd5-8738-c785df002c93" (UID: "0e54e31f-e9f8-4dd5-8738-c785df002c93"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.683015 4813 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.683039 4813 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0e54e31f-e9f8-4dd5-8738-c785df002c93-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.683057 4813 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.683075 4813 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.683092 4813 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.683110 4813 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.683128 4813 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0e54e31f-e9f8-4dd5-8738-c785df002c93-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.690649 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e54e31f-e9f8-4dd5-8738-c785df002c93-kube-api-access-nbvf9" (OuterVolumeSpecName: "kube-api-access-nbvf9") pod "0e54e31f-e9f8-4dd5-8738-c785df002c93" (UID: "0e54e31f-e9f8-4dd5-8738-c785df002c93"). InnerVolumeSpecName "kube-api-access-nbvf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.692480 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e54e31f-e9f8-4dd5-8738-c785df002c93-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "0e54e31f-e9f8-4dd5-8738-c785df002c93" (UID: "0e54e31f-e9f8-4dd5-8738-c785df002c93"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.711115 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "0e54e31f-e9f8-4dd5-8738-c785df002c93" (UID: "0e54e31f-e9f8-4dd5-8738-c785df002c93"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.784488 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/52f40cf1-858f-4e11-867f-9119a513a03a-ovnkube-config\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.784702 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-systemd-units\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.784797 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/52f40cf1-858f-4e11-867f-9119a513a03a-ovnkube-script-lib\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.784871 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-node-log\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.784939 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/52f40cf1-858f-4e11-867f-9119a513a03a-ovn-node-metrics-cert\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.785008 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9p7s\" (UniqueName: \"kubernetes.io/projected/52f40cf1-858f-4e11-867f-9119a513a03a-kube-api-access-q9p7s\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.785090 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-host-slash\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.785169 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-host-cni-bin\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.785237 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-host-run-netns\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.785320 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-host-run-ovn-kubernetes\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.785387 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-host-kubelet\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.785450 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-run-ovn\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.785527 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-var-lib-openvswitch\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.785596 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-host-cni-netd\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.785667 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-run-openvswitch\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.785733 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-log-socket\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.785806 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-etc-openvswitch\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.785874 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/52f40cf1-858f-4e11-867f-9119a513a03a-env-overrides\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.785952 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.786045 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-run-systemd\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.786135 4813 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0e54e31f-e9f8-4dd5-8738-c785df002c93-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.786194 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbvf9\" (UniqueName: \"kubernetes.io/projected/0e54e31f-e9f8-4dd5-8738-c785df002c93-kube-api-access-nbvf9\") on node \"crc\" DevicePath \"\"" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.786243 4813 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0e54e31f-e9f8-4dd5-8738-c785df002c93-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.786316 4813 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0e54e31f-e9f8-4dd5-8738-c785df002c93-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.887493 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-host-kubelet\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.887579 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-run-ovn\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.887616 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-var-lib-openvswitch\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.887655 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-host-cni-netd\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.887690 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-run-openvswitch\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.887728 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-log-socket\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.887769 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-etc-openvswitch\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.887797 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/52f40cf1-858f-4e11-867f-9119a513a03a-env-overrides\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.887851 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.887893 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-run-systemd\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.887940 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/52f40cf1-858f-4e11-867f-9119a513a03a-ovnkube-config\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.887978 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-systemd-units\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.888006 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/52f40cf1-858f-4e11-867f-9119a513a03a-ovnkube-script-lib\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.888035 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-node-log\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.888068 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/52f40cf1-858f-4e11-867f-9119a513a03a-ovn-node-metrics-cert\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.888103 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9p7s\" (UniqueName: \"kubernetes.io/projected/52f40cf1-858f-4e11-867f-9119a513a03a-kube-api-access-q9p7s\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.888135 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-host-slash\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.888214 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-host-cni-bin\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.888244 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-host-run-netns\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.888297 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-host-run-ovn-kubernetes\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.888406 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-host-run-ovn-kubernetes\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.888467 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-host-kubelet\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.888505 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-run-ovn\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.888548 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-var-lib-openvswitch\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.888625 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-host-cni-netd\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.888663 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-run-openvswitch\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.888713 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-log-socket\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.888760 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-etc-openvswitch\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.889440 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-host-slash\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.889442 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-node-log\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.889531 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-host-cni-bin\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.889569 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-host-run-netns\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.889759 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/52f40cf1-858f-4e11-867f-9119a513a03a-env-overrides\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.889829 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-systemd-units\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.889952 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-run-systemd\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.890129 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/52f40cf1-858f-4e11-867f-9119a513a03a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.890233 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/52f40cf1-858f-4e11-867f-9119a513a03a-ovnkube-config\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.890740 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/52f40cf1-858f-4e11-867f-9119a513a03a-ovnkube-script-lib\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.893589 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/52f40cf1-858f-4e11-867f-9119a513a03a-ovn-node-metrics-cert\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:09 crc kubenswrapper[4813]: I1206 15:56:09.918815 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9p7s\" (UniqueName: \"kubernetes.io/projected/52f40cf1-858f-4e11-867f-9119a513a03a-kube-api-access-q9p7s\") pod \"ovnkube-node-lnmrs\" (UID: \"52f40cf1-858f-4e11-867f-9119a513a03a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.006782 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:10 crc kubenswrapper[4813]: W1206 15:56:10.038775 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52f40cf1_858f_4e11_867f_9119a513a03a.slice/crio-15150a9de229400b815b50a9efea6c1e82b45a0b58004b2ae3cd973028777ab2 WatchSource:0}: Error finding container 15150a9de229400b815b50a9efea6c1e82b45a0b58004b2ae3cd973028777ab2: Status 404 returned error can't find the container with id 15150a9de229400b815b50a9efea6c1e82b45a0b58004b2ae3cd973028777ab2 Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.236569 4813 generic.go:334] "Generic (PLEG): container finished" podID="52f40cf1-858f-4e11-867f-9119a513a03a" containerID="04e9b012d91050575646992a3513f7b6f95e836a4af770637d10674c19e7341a" exitCode=0 Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.236687 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" event={"ID":"52f40cf1-858f-4e11-867f-9119a513a03a","Type":"ContainerDied","Data":"04e9b012d91050575646992a3513f7b6f95e836a4af770637d10674c19e7341a"} Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.236736 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" event={"ID":"52f40cf1-858f-4e11-867f-9119a513a03a","Type":"ContainerStarted","Data":"15150a9de229400b815b50a9efea6c1e82b45a0b58004b2ae3cd973028777ab2"} Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.249774 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xdt4d_0e54e31f-e9f8-4dd5-8738-c785df002c93/ovn-acl-logging/0.log" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.250666 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xdt4d_0e54e31f-e9f8-4dd5-8738-c785df002c93/ovn-controller/0.log" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.251467 4813 generic.go:334] "Generic (PLEG): container finished" podID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerID="afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2" exitCode=0 Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.251720 4813 generic.go:334] "Generic (PLEG): container finished" podID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerID="2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45" exitCode=0 Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.251976 4813 generic.go:334] "Generic (PLEG): container finished" podID="0e54e31f-e9f8-4dd5-8738-c785df002c93" containerID="57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc" exitCode=0 Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.251831 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.251797 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" event={"ID":"0e54e31f-e9f8-4dd5-8738-c785df002c93","Type":"ContainerDied","Data":"afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2"} Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.252489 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" event={"ID":"0e54e31f-e9f8-4dd5-8738-c785df002c93","Type":"ContainerDied","Data":"2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45"} Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.252521 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" event={"ID":"0e54e31f-e9f8-4dd5-8738-c785df002c93","Type":"ContainerDied","Data":"57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc"} Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.252543 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xdt4d" event={"ID":"0e54e31f-e9f8-4dd5-8738-c785df002c93","Type":"ContainerDied","Data":"4ab555fa8c335cb15fd9727b52dd5a7cb19d2550221f6cd57f88520d7f2e72a8"} Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.252563 4813 scope.go:117] "RemoveContainer" containerID="5cf32df4a99c42b7b090412102796d4a6d265bc5a4de2ecf1648cadbac28ea73" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.269756 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gl495_856bb45b-525c-4ef0-bf7c-0691cf54b342/kube-multus/2.log" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.302144 4813 scope.go:117] "RemoveContainer" containerID="afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.325091 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-xdt4d"] Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.333582 4813 scope.go:117] "RemoveContainer" containerID="2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.336187 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-xdt4d"] Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.360214 4813 scope.go:117] "RemoveContainer" containerID="57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.373841 4813 scope.go:117] "RemoveContainer" containerID="2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.387375 4813 scope.go:117] "RemoveContainer" containerID="6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.399381 4813 scope.go:117] "RemoveContainer" containerID="f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.411607 4813 scope.go:117] "RemoveContainer" containerID="5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.429365 4813 scope.go:117] "RemoveContainer" containerID="4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.442650 4813 scope.go:117] "RemoveContainer" containerID="5cf32df4a99c42b7b090412102796d4a6d265bc5a4de2ecf1648cadbac28ea73" Dec 06 15:56:10 crc kubenswrapper[4813]: E1206 15:56:10.443214 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cf32df4a99c42b7b090412102796d4a6d265bc5a4de2ecf1648cadbac28ea73\": container with ID starting with 5cf32df4a99c42b7b090412102796d4a6d265bc5a4de2ecf1648cadbac28ea73 not found: ID does not exist" containerID="5cf32df4a99c42b7b090412102796d4a6d265bc5a4de2ecf1648cadbac28ea73" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.443293 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cf32df4a99c42b7b090412102796d4a6d265bc5a4de2ecf1648cadbac28ea73"} err="failed to get container status \"5cf32df4a99c42b7b090412102796d4a6d265bc5a4de2ecf1648cadbac28ea73\": rpc error: code = NotFound desc = could not find container \"5cf32df4a99c42b7b090412102796d4a6d265bc5a4de2ecf1648cadbac28ea73\": container with ID starting with 5cf32df4a99c42b7b090412102796d4a6d265bc5a4de2ecf1648cadbac28ea73 not found: ID does not exist" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.443324 4813 scope.go:117] "RemoveContainer" containerID="afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2" Dec 06 15:56:10 crc kubenswrapper[4813]: E1206 15:56:10.443705 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\": container with ID starting with afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2 not found: ID does not exist" containerID="afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.443745 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2"} err="failed to get container status \"afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\": rpc error: code = NotFound desc = could not find container \"afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\": container with ID starting with afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2 not found: ID does not exist" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.443772 4813 scope.go:117] "RemoveContainer" containerID="2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45" Dec 06 15:56:10 crc kubenswrapper[4813]: E1206 15:56:10.444079 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\": container with ID starting with 2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45 not found: ID does not exist" containerID="2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.444100 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45"} err="failed to get container status \"2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\": rpc error: code = NotFound desc = could not find container \"2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\": container with ID starting with 2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45 not found: ID does not exist" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.444115 4813 scope.go:117] "RemoveContainer" containerID="57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc" Dec 06 15:56:10 crc kubenswrapper[4813]: E1206 15:56:10.444738 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\": container with ID starting with 57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc not found: ID does not exist" containerID="57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.444890 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc"} err="failed to get container status \"57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\": rpc error: code = NotFound desc = could not find container \"57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\": container with ID starting with 57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc not found: ID does not exist" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.444988 4813 scope.go:117] "RemoveContainer" containerID="2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa" Dec 06 15:56:10 crc kubenswrapper[4813]: E1206 15:56:10.445624 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\": container with ID starting with 2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa not found: ID does not exist" containerID="2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.445651 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa"} err="failed to get container status \"2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\": rpc error: code = NotFound desc = could not find container \"2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\": container with ID starting with 2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa not found: ID does not exist" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.445665 4813 scope.go:117] "RemoveContainer" containerID="6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e" Dec 06 15:56:10 crc kubenswrapper[4813]: E1206 15:56:10.446069 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\": container with ID starting with 6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e not found: ID does not exist" containerID="6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.446158 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e"} err="failed to get container status \"6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\": rpc error: code = NotFound desc = could not find container \"6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\": container with ID starting with 6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e not found: ID does not exist" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.446221 4813 scope.go:117] "RemoveContainer" containerID="f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c" Dec 06 15:56:10 crc kubenswrapper[4813]: E1206 15:56:10.446651 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\": container with ID starting with f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c not found: ID does not exist" containerID="f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.446682 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c"} err="failed to get container status \"f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\": rpc error: code = NotFound desc = could not find container \"f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\": container with ID starting with f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c not found: ID does not exist" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.446701 4813 scope.go:117] "RemoveContainer" containerID="5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c" Dec 06 15:56:10 crc kubenswrapper[4813]: E1206 15:56:10.447123 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\": container with ID starting with 5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c not found: ID does not exist" containerID="5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.447205 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c"} err="failed to get container status \"5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\": rpc error: code = NotFound desc = could not find container \"5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\": container with ID starting with 5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c not found: ID does not exist" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.447302 4813 scope.go:117] "RemoveContainer" containerID="4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014" Dec 06 15:56:10 crc kubenswrapper[4813]: E1206 15:56:10.447644 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\": container with ID starting with 4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014 not found: ID does not exist" containerID="4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.447672 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014"} err="failed to get container status \"4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\": rpc error: code = NotFound desc = could not find container \"4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\": container with ID starting with 4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014 not found: ID does not exist" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.447692 4813 scope.go:117] "RemoveContainer" containerID="5cf32df4a99c42b7b090412102796d4a6d265bc5a4de2ecf1648cadbac28ea73" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.447890 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cf32df4a99c42b7b090412102796d4a6d265bc5a4de2ecf1648cadbac28ea73"} err="failed to get container status \"5cf32df4a99c42b7b090412102796d4a6d265bc5a4de2ecf1648cadbac28ea73\": rpc error: code = NotFound desc = could not find container \"5cf32df4a99c42b7b090412102796d4a6d265bc5a4de2ecf1648cadbac28ea73\": container with ID starting with 5cf32df4a99c42b7b090412102796d4a6d265bc5a4de2ecf1648cadbac28ea73 not found: ID does not exist" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.447911 4813 scope.go:117] "RemoveContainer" containerID="afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.448156 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2"} err="failed to get container status \"afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\": rpc error: code = NotFound desc = could not find container \"afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\": container with ID starting with afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2 not found: ID does not exist" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.448181 4813 scope.go:117] "RemoveContainer" containerID="2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.448521 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45"} err="failed to get container status \"2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\": rpc error: code = NotFound desc = could not find container \"2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\": container with ID starting with 2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45 not found: ID does not exist" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.448602 4813 scope.go:117] "RemoveContainer" containerID="57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.448922 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc"} err="failed to get container status \"57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\": rpc error: code = NotFound desc = could not find container \"57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\": container with ID starting with 57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc not found: ID does not exist" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.448947 4813 scope.go:117] "RemoveContainer" containerID="2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.449145 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa"} err="failed to get container status \"2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\": rpc error: code = NotFound desc = could not find container \"2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\": container with ID starting with 2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa not found: ID does not exist" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.449175 4813 scope.go:117] "RemoveContainer" containerID="6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.449410 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e"} err="failed to get container status \"6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\": rpc error: code = NotFound desc = could not find container \"6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\": container with ID starting with 6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e not found: ID does not exist" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.449433 4813 scope.go:117] "RemoveContainer" containerID="f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.449625 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c"} err="failed to get container status \"f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\": rpc error: code = NotFound desc = could not find container \"f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\": container with ID starting with f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c not found: ID does not exist" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.449660 4813 scope.go:117] "RemoveContainer" containerID="5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.449883 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c"} err="failed to get container status \"5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\": rpc error: code = NotFound desc = could not find container \"5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\": container with ID starting with 5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c not found: ID does not exist" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.449907 4813 scope.go:117] "RemoveContainer" containerID="4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.450137 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014"} err="failed to get container status \"4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\": rpc error: code = NotFound desc = could not find container \"4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\": container with ID starting with 4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014 not found: ID does not exist" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.450164 4813 scope.go:117] "RemoveContainer" containerID="5cf32df4a99c42b7b090412102796d4a6d265bc5a4de2ecf1648cadbac28ea73" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.450372 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cf32df4a99c42b7b090412102796d4a6d265bc5a4de2ecf1648cadbac28ea73"} err="failed to get container status \"5cf32df4a99c42b7b090412102796d4a6d265bc5a4de2ecf1648cadbac28ea73\": rpc error: code = NotFound desc = could not find container \"5cf32df4a99c42b7b090412102796d4a6d265bc5a4de2ecf1648cadbac28ea73\": container with ID starting with 5cf32df4a99c42b7b090412102796d4a6d265bc5a4de2ecf1648cadbac28ea73 not found: ID does not exist" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.450402 4813 scope.go:117] "RemoveContainer" containerID="afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.450591 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2"} err="failed to get container status \"afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\": rpc error: code = NotFound desc = could not find container \"afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2\": container with ID starting with afd131c2a7d2d8aa93c27428c9f152f8e95a7873ec18496c2ef747b4f58f89e2 not found: ID does not exist" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.450629 4813 scope.go:117] "RemoveContainer" containerID="2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.450801 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45"} err="failed to get container status \"2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\": rpc error: code = NotFound desc = could not find container \"2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45\": container with ID starting with 2ba5963725f1eed86c339658e57a0535efb1d04c54060192ab2c4df1d72c0f45 not found: ID does not exist" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.450823 4813 scope.go:117] "RemoveContainer" containerID="57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.450986 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc"} err="failed to get container status \"57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\": rpc error: code = NotFound desc = could not find container \"57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc\": container with ID starting with 57f9cf3b0d0b67e9d954d24124505e4e57e1333bcb92483b5225802dd46698fc not found: ID does not exist" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.451010 4813 scope.go:117] "RemoveContainer" containerID="2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.451171 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa"} err="failed to get container status \"2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\": rpc error: code = NotFound desc = could not find container \"2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa\": container with ID starting with 2190f9d909d3eefaa3c3f9e6f51b564610835d28b88ca984ee64f7d8fdf325fa not found: ID does not exist" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.451192 4813 scope.go:117] "RemoveContainer" containerID="6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.451404 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e"} err="failed to get container status \"6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\": rpc error: code = NotFound desc = could not find container \"6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e\": container with ID starting with 6d6a0fe317b45f9f2d4c0d1494c536a20f480edff0bf431ddb03aea588ba677e not found: ID does not exist" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.451425 4813 scope.go:117] "RemoveContainer" containerID="f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.451617 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c"} err="failed to get container status \"f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\": rpc error: code = NotFound desc = could not find container \"f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c\": container with ID starting with f39edd32bebad2c9c99dbad0ebcddac2bbc13cd84cd4d0cc03f28f8bffc3f81c not found: ID does not exist" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.451634 4813 scope.go:117] "RemoveContainer" containerID="5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.451825 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c"} err="failed to get container status \"5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\": rpc error: code = NotFound desc = could not find container \"5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c\": container with ID starting with 5d71cc90a7b5ed655883f87502201e633570cc0f178074fd24772ce176d4ce9c not found: ID does not exist" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.451842 4813 scope.go:117] "RemoveContainer" containerID="4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.452013 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014"} err="failed to get container status \"4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\": rpc error: code = NotFound desc = could not find container \"4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014\": container with ID starting with 4d89c1b0a6d0d6bdbc657d8523949e96a6352bd85ad4f66b91f7ebdf82d4a014 not found: ID does not exist" Dec 06 15:56:10 crc kubenswrapper[4813]: I1206 15:56:10.495477 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e54e31f-e9f8-4dd5-8738-c785df002c93" path="/var/lib/kubelet/pods/0e54e31f-e9f8-4dd5-8738-c785df002c93/volumes" Dec 06 15:56:11 crc kubenswrapper[4813]: I1206 15:56:11.280986 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" event={"ID":"52f40cf1-858f-4e11-867f-9119a513a03a","Type":"ContainerStarted","Data":"414b2858633e3ea8facc85e1076ea7e03bd855214bfc7a67eac0a6ea0b08abbc"} Dec 06 15:56:11 crc kubenswrapper[4813]: I1206 15:56:11.281075 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" event={"ID":"52f40cf1-858f-4e11-867f-9119a513a03a","Type":"ContainerStarted","Data":"18cc058e38938e0d211543a8499a15372eed2cbe67fff7589c3ec77dc7f04372"} Dec 06 15:56:11 crc kubenswrapper[4813]: I1206 15:56:11.281103 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" event={"ID":"52f40cf1-858f-4e11-867f-9119a513a03a","Type":"ContainerStarted","Data":"15061946483a1b1b45ed1cfdf699ac68914e10f76b4f0e14eb0c7189b645594d"} Dec 06 15:56:11 crc kubenswrapper[4813]: I1206 15:56:11.281124 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" event={"ID":"52f40cf1-858f-4e11-867f-9119a513a03a","Type":"ContainerStarted","Data":"2baa329f9e14de5baa586f4f1ba0590d5fdf746b8d731ef21459c79913dc29ab"} Dec 06 15:56:11 crc kubenswrapper[4813]: I1206 15:56:11.281145 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" event={"ID":"52f40cf1-858f-4e11-867f-9119a513a03a","Type":"ContainerStarted","Data":"42e7f82b9493f15a45217f1d8893f4d6c77dd6d51a38ffddce623b413259b31a"} Dec 06 15:56:11 crc kubenswrapper[4813]: I1206 15:56:11.281171 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" event={"ID":"52f40cf1-858f-4e11-867f-9119a513a03a","Type":"ContainerStarted","Data":"710fa5c73e7b1be31346003f2d1324dd23788f995a454c3e816e57a021f97001"} Dec 06 15:56:13 crc kubenswrapper[4813]: I1206 15:56:13.303947 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" event={"ID":"52f40cf1-858f-4e11-867f-9119a513a03a","Type":"ContainerStarted","Data":"83940338c12cc622719e076a74ea5423a370c06d14d3c5ad1f919ce092e4dd8d"} Dec 06 15:56:16 crc kubenswrapper[4813]: I1206 15:56:16.342857 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" event={"ID":"52f40cf1-858f-4e11-867f-9119a513a03a","Type":"ContainerStarted","Data":"9dfb6234fd3d4c5d5675888b8dd711c2012833d550b4f600ddeb28809261e47d"} Dec 06 15:56:16 crc kubenswrapper[4813]: I1206 15:56:16.343658 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:16 crc kubenswrapper[4813]: I1206 15:56:16.402167 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:16 crc kubenswrapper[4813]: I1206 15:56:16.414375 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" podStartSLOduration=7.414356609 podStartE2EDuration="7.414356609s" podCreationTimestamp="2025-12-06 15:56:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:56:16.410071103 +0000 UTC m=+616.300950679" watchObservedRunningTime="2025-12-06 15:56:16.414356609 +0000 UTC m=+616.305236195" Dec 06 15:56:17 crc kubenswrapper[4813]: I1206 15:56:17.350329 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:17 crc kubenswrapper[4813]: I1206 15:56:17.350756 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:17 crc kubenswrapper[4813]: I1206 15:56:17.382910 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:19 crc kubenswrapper[4813]: I1206 15:56:19.427777 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 15:56:19 crc kubenswrapper[4813]: I1206 15:56:19.427855 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 15:56:19 crc kubenswrapper[4813]: I1206 15:56:19.427905 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" Dec 06 15:56:19 crc kubenswrapper[4813]: I1206 15:56:19.428867 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6841d4ac1913266f58304ddc35c3d4b57d8eed36652e13d503003868ba785be0"} pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 15:56:19 crc kubenswrapper[4813]: I1206 15:56:19.428955 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" containerID="cri-o://6841d4ac1913266f58304ddc35c3d4b57d8eed36652e13d503003868ba785be0" gracePeriod=600 Dec 06 15:56:20 crc kubenswrapper[4813]: I1206 15:56:20.391178 4813 generic.go:334] "Generic (PLEG): container finished" podID="d88e8bae-c055-4c55-b548-f621ff96de06" containerID="6841d4ac1913266f58304ddc35c3d4b57d8eed36652e13d503003868ba785be0" exitCode=0 Dec 06 15:56:20 crc kubenswrapper[4813]: I1206 15:56:20.391248 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerDied","Data":"6841d4ac1913266f58304ddc35c3d4b57d8eed36652e13d503003868ba785be0"} Dec 06 15:56:20 crc kubenswrapper[4813]: I1206 15:56:20.391314 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerStarted","Data":"0d0cad4af55e6ecbb2a20d8a543c0848e24b8077a79dec0b4034636d4a17f514"} Dec 06 15:56:20 crc kubenswrapper[4813]: I1206 15:56:20.391342 4813 scope.go:117] "RemoveContainer" containerID="9bb3ab0a5b69e0227b05bdb28844955eb8df2d6874e1937ca3d3672f21c852e9" Dec 06 15:56:21 crc kubenswrapper[4813]: I1206 15:56:21.487438 4813 scope.go:117] "RemoveContainer" containerID="ef40f8e7e7629e97c914ef68c7d15da2e144cc9558102b30198efb0618912fe8" Dec 06 15:56:21 crc kubenswrapper[4813]: E1206 15:56:21.488025 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-gl495_openshift-multus(856bb45b-525c-4ef0-bf7c-0691cf54b342)\"" pod="openshift-multus/multus-gl495" podUID="856bb45b-525c-4ef0-bf7c-0691cf54b342" Dec 06 15:56:34 crc kubenswrapper[4813]: I1206 15:56:34.487113 4813 scope.go:117] "RemoveContainer" containerID="ef40f8e7e7629e97c914ef68c7d15da2e144cc9558102b30198efb0618912fe8" Dec 06 15:56:35 crc kubenswrapper[4813]: I1206 15:56:35.512197 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gl495_856bb45b-525c-4ef0-bf7c-0691cf54b342/kube-multus/2.log" Dec 06 15:56:35 crc kubenswrapper[4813]: I1206 15:56:35.513611 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gl495" event={"ID":"856bb45b-525c-4ef0-bf7c-0691cf54b342","Type":"ContainerStarted","Data":"6a8032509cd0e1f75ece0ca045d78b5d0ce0f28a718ce30ca97c76b24ff82cec"} Dec 06 15:56:40 crc kubenswrapper[4813]: I1206 15:56:40.049058 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lnmrs" Dec 06 15:56:49 crc kubenswrapper[4813]: I1206 15:56:49.266890 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx"] Dec 06 15:56:49 crc kubenswrapper[4813]: I1206 15:56:49.268616 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx" Dec 06 15:56:49 crc kubenswrapper[4813]: I1206 15:56:49.276798 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 06 15:56:49 crc kubenswrapper[4813]: I1206 15:56:49.285495 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx"] Dec 06 15:56:49 crc kubenswrapper[4813]: I1206 15:56:49.336435 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f977g\" (UniqueName: \"kubernetes.io/projected/4cc38736-85b3-4e2d-87fe-604eebf52317-kube-api-access-f977g\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx\" (UID: \"4cc38736-85b3-4e2d-87fe-604eebf52317\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx" Dec 06 15:56:49 crc kubenswrapper[4813]: I1206 15:56:49.336615 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4cc38736-85b3-4e2d-87fe-604eebf52317-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx\" (UID: \"4cc38736-85b3-4e2d-87fe-604eebf52317\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx" Dec 06 15:56:49 crc kubenswrapper[4813]: I1206 15:56:49.336691 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4cc38736-85b3-4e2d-87fe-604eebf52317-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx\" (UID: \"4cc38736-85b3-4e2d-87fe-604eebf52317\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx" Dec 06 15:56:49 crc kubenswrapper[4813]: I1206 15:56:49.437676 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4cc38736-85b3-4e2d-87fe-604eebf52317-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx\" (UID: \"4cc38736-85b3-4e2d-87fe-604eebf52317\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx" Dec 06 15:56:49 crc kubenswrapper[4813]: I1206 15:56:49.437740 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f977g\" (UniqueName: \"kubernetes.io/projected/4cc38736-85b3-4e2d-87fe-604eebf52317-kube-api-access-f977g\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx\" (UID: \"4cc38736-85b3-4e2d-87fe-604eebf52317\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx" Dec 06 15:56:49 crc kubenswrapper[4813]: I1206 15:56:49.437839 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4cc38736-85b3-4e2d-87fe-604eebf52317-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx\" (UID: \"4cc38736-85b3-4e2d-87fe-604eebf52317\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx" Dec 06 15:56:49 crc kubenswrapper[4813]: I1206 15:56:49.438535 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4cc38736-85b3-4e2d-87fe-604eebf52317-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx\" (UID: \"4cc38736-85b3-4e2d-87fe-604eebf52317\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx" Dec 06 15:56:49 crc kubenswrapper[4813]: I1206 15:56:49.438532 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4cc38736-85b3-4e2d-87fe-604eebf52317-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx\" (UID: \"4cc38736-85b3-4e2d-87fe-604eebf52317\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx" Dec 06 15:56:49 crc kubenswrapper[4813]: I1206 15:56:49.470254 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f977g\" (UniqueName: \"kubernetes.io/projected/4cc38736-85b3-4e2d-87fe-604eebf52317-kube-api-access-f977g\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx\" (UID: \"4cc38736-85b3-4e2d-87fe-604eebf52317\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx" Dec 06 15:56:49 crc kubenswrapper[4813]: I1206 15:56:49.614951 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx" Dec 06 15:56:49 crc kubenswrapper[4813]: I1206 15:56:49.905586 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx"] Dec 06 15:56:50 crc kubenswrapper[4813]: I1206 15:56:50.615694 4813 generic.go:334] "Generic (PLEG): container finished" podID="4cc38736-85b3-4e2d-87fe-604eebf52317" containerID="662e4b91fa954eb5e50b2c959e758f13a3e9c1ed5545a74952fa769ff1e982fa" exitCode=0 Dec 06 15:56:50 crc kubenswrapper[4813]: I1206 15:56:50.616082 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx" event={"ID":"4cc38736-85b3-4e2d-87fe-604eebf52317","Type":"ContainerDied","Data":"662e4b91fa954eb5e50b2c959e758f13a3e9c1ed5545a74952fa769ff1e982fa"} Dec 06 15:56:50 crc kubenswrapper[4813]: I1206 15:56:50.616130 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx" event={"ID":"4cc38736-85b3-4e2d-87fe-604eebf52317","Type":"ContainerStarted","Data":"e32f7cf043ac326cbd80af287222a04e163db869df817fa4c006b11c9c7ccff6"} Dec 06 15:56:51 crc kubenswrapper[4813]: I1206 15:56:51.623026 4813 generic.go:334] "Generic (PLEG): container finished" podID="4cc38736-85b3-4e2d-87fe-604eebf52317" containerID="328c31ea584586b19b07072982db5c773009632407ff66b5ba4a75afce4bb65e" exitCode=0 Dec 06 15:56:51 crc kubenswrapper[4813]: I1206 15:56:51.623122 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx" event={"ID":"4cc38736-85b3-4e2d-87fe-604eebf52317","Type":"ContainerDied","Data":"328c31ea584586b19b07072982db5c773009632407ff66b5ba4a75afce4bb65e"} Dec 06 15:56:52 crc kubenswrapper[4813]: I1206 15:56:52.635154 4813 generic.go:334] "Generic (PLEG): container finished" podID="4cc38736-85b3-4e2d-87fe-604eebf52317" containerID="1e25ec14db0cae40ded4608d67d22a293c2ee5dd9e9465e547dfe18b31e3c93d" exitCode=0 Dec 06 15:56:52 crc kubenswrapper[4813]: I1206 15:56:52.635227 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx" event={"ID":"4cc38736-85b3-4e2d-87fe-604eebf52317","Type":"ContainerDied","Data":"1e25ec14db0cae40ded4608d67d22a293c2ee5dd9e9465e547dfe18b31e3c93d"} Dec 06 15:56:54 crc kubenswrapper[4813]: I1206 15:56:54.315285 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx" Dec 06 15:56:54 crc kubenswrapper[4813]: I1206 15:56:54.501881 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4cc38736-85b3-4e2d-87fe-604eebf52317-util\") pod \"4cc38736-85b3-4e2d-87fe-604eebf52317\" (UID: \"4cc38736-85b3-4e2d-87fe-604eebf52317\") " Dec 06 15:56:54 crc kubenswrapper[4813]: I1206 15:56:54.501959 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4cc38736-85b3-4e2d-87fe-604eebf52317-bundle\") pod \"4cc38736-85b3-4e2d-87fe-604eebf52317\" (UID: \"4cc38736-85b3-4e2d-87fe-604eebf52317\") " Dec 06 15:56:54 crc kubenswrapper[4813]: I1206 15:56:54.502029 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f977g\" (UniqueName: \"kubernetes.io/projected/4cc38736-85b3-4e2d-87fe-604eebf52317-kube-api-access-f977g\") pod \"4cc38736-85b3-4e2d-87fe-604eebf52317\" (UID: \"4cc38736-85b3-4e2d-87fe-604eebf52317\") " Dec 06 15:56:54 crc kubenswrapper[4813]: I1206 15:56:54.502657 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cc38736-85b3-4e2d-87fe-604eebf52317-bundle" (OuterVolumeSpecName: "bundle") pod "4cc38736-85b3-4e2d-87fe-604eebf52317" (UID: "4cc38736-85b3-4e2d-87fe-604eebf52317"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:56:54 crc kubenswrapper[4813]: I1206 15:56:54.510468 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cc38736-85b3-4e2d-87fe-604eebf52317-kube-api-access-f977g" (OuterVolumeSpecName: "kube-api-access-f977g") pod "4cc38736-85b3-4e2d-87fe-604eebf52317" (UID: "4cc38736-85b3-4e2d-87fe-604eebf52317"). InnerVolumeSpecName "kube-api-access-f977g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:56:54 crc kubenswrapper[4813]: I1206 15:56:54.530175 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cc38736-85b3-4e2d-87fe-604eebf52317-util" (OuterVolumeSpecName: "util") pod "4cc38736-85b3-4e2d-87fe-604eebf52317" (UID: "4cc38736-85b3-4e2d-87fe-604eebf52317"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:56:54 crc kubenswrapper[4813]: I1206 15:56:54.604359 4813 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4cc38736-85b3-4e2d-87fe-604eebf52317-util\") on node \"crc\" DevicePath \"\"" Dec 06 15:56:54 crc kubenswrapper[4813]: I1206 15:56:54.604704 4813 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4cc38736-85b3-4e2d-87fe-604eebf52317-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 15:56:54 crc kubenswrapper[4813]: I1206 15:56:54.604759 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f977g\" (UniqueName: \"kubernetes.io/projected/4cc38736-85b3-4e2d-87fe-604eebf52317-kube-api-access-f977g\") on node \"crc\" DevicePath \"\"" Dec 06 15:56:54 crc kubenswrapper[4813]: I1206 15:56:54.649482 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx" event={"ID":"4cc38736-85b3-4e2d-87fe-604eebf52317","Type":"ContainerDied","Data":"e32f7cf043ac326cbd80af287222a04e163db869df817fa4c006b11c9c7ccff6"} Dec 06 15:56:54 crc kubenswrapper[4813]: I1206 15:56:54.649571 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx" Dec 06 15:56:54 crc kubenswrapper[4813]: I1206 15:56:54.649594 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e32f7cf043ac326cbd80af287222a04e163db869df817fa4c006b11c9c7ccff6" Dec 06 15:56:56 crc kubenswrapper[4813]: I1206 15:56:56.337162 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-bd2wf"] Dec 06 15:56:56 crc kubenswrapper[4813]: E1206 15:56:56.337395 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cc38736-85b3-4e2d-87fe-604eebf52317" containerName="extract" Dec 06 15:56:56 crc kubenswrapper[4813]: I1206 15:56:56.337409 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cc38736-85b3-4e2d-87fe-604eebf52317" containerName="extract" Dec 06 15:56:56 crc kubenswrapper[4813]: E1206 15:56:56.337422 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cc38736-85b3-4e2d-87fe-604eebf52317" containerName="pull" Dec 06 15:56:56 crc kubenswrapper[4813]: I1206 15:56:56.337430 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cc38736-85b3-4e2d-87fe-604eebf52317" containerName="pull" Dec 06 15:56:56 crc kubenswrapper[4813]: E1206 15:56:56.337444 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cc38736-85b3-4e2d-87fe-604eebf52317" containerName="util" Dec 06 15:56:56 crc kubenswrapper[4813]: I1206 15:56:56.337452 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cc38736-85b3-4e2d-87fe-604eebf52317" containerName="util" Dec 06 15:56:56 crc kubenswrapper[4813]: I1206 15:56:56.337565 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cc38736-85b3-4e2d-87fe-604eebf52317" containerName="extract" Dec 06 15:56:56 crc kubenswrapper[4813]: I1206 15:56:56.337973 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bd2wf" Dec 06 15:56:56 crc kubenswrapper[4813]: I1206 15:56:56.342311 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 06 15:56:56 crc kubenswrapper[4813]: I1206 15:56:56.342560 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 06 15:56:56 crc kubenswrapper[4813]: I1206 15:56:56.346528 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-bd2wf"] Dec 06 15:56:56 crc kubenswrapper[4813]: I1206 15:56:56.354161 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-xhzgm" Dec 06 15:56:56 crc kubenswrapper[4813]: I1206 15:56:56.432841 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk8v7\" (UniqueName: \"kubernetes.io/projected/215ac015-9856-4c78-ab4d-0035ee29b678-kube-api-access-nk8v7\") pod \"nmstate-operator-5b5b58f5c8-bd2wf\" (UID: \"215ac015-9856-4c78-ab4d-0035ee29b678\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bd2wf" Dec 06 15:56:56 crc kubenswrapper[4813]: I1206 15:56:56.534327 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk8v7\" (UniqueName: \"kubernetes.io/projected/215ac015-9856-4c78-ab4d-0035ee29b678-kube-api-access-nk8v7\") pod \"nmstate-operator-5b5b58f5c8-bd2wf\" (UID: \"215ac015-9856-4c78-ab4d-0035ee29b678\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bd2wf" Dec 06 15:56:56 crc kubenswrapper[4813]: I1206 15:56:56.569794 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk8v7\" (UniqueName: \"kubernetes.io/projected/215ac015-9856-4c78-ab4d-0035ee29b678-kube-api-access-nk8v7\") pod \"nmstate-operator-5b5b58f5c8-bd2wf\" (UID: \"215ac015-9856-4c78-ab4d-0035ee29b678\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bd2wf" Dec 06 15:56:56 crc kubenswrapper[4813]: I1206 15:56:56.653985 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bd2wf" Dec 06 15:56:57 crc kubenswrapper[4813]: I1206 15:56:57.090410 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-bd2wf"] Dec 06 15:56:57 crc kubenswrapper[4813]: I1206 15:56:57.668979 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bd2wf" event={"ID":"215ac015-9856-4c78-ab4d-0035ee29b678","Type":"ContainerStarted","Data":"454eb8b26d52caf5c52b9933f1a71b6bbbd90fe68893b253c5fffd57b611933f"} Dec 06 15:56:59 crc kubenswrapper[4813]: I1206 15:56:59.680040 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bd2wf" event={"ID":"215ac015-9856-4c78-ab4d-0035ee29b678","Type":"ContainerStarted","Data":"25414c11516d24b46a4258ce4192a4655af6610e6dccd0b74999361d1ab7a148"} Dec 06 15:56:59 crc kubenswrapper[4813]: I1206 15:56:59.704015 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bd2wf" podStartSLOduration=1.815116644 podStartE2EDuration="3.703992577s" podCreationTimestamp="2025-12-06 15:56:56 +0000 UTC" firstStartedPulling="2025-12-06 15:56:57.101359826 +0000 UTC m=+656.992239402" lastFinishedPulling="2025-12-06 15:56:58.990235759 +0000 UTC m=+658.881115335" observedRunningTime="2025-12-06 15:56:59.702930089 +0000 UTC m=+659.593809705" watchObservedRunningTime="2025-12-06 15:56:59.703992577 +0000 UTC m=+659.594872183" Dec 06 15:57:00 crc kubenswrapper[4813]: I1206 15:57:00.663915 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-f4x5w"] Dec 06 15:57:00 crc kubenswrapper[4813]: I1206 15:57:00.664973 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-f4x5w" Dec 06 15:57:00 crc kubenswrapper[4813]: I1206 15:57:00.687825 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vc5f4"] Dec 06 15:57:00 crc kubenswrapper[4813]: I1206 15:57:00.689106 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vc5f4" Dec 06 15:57:00 crc kubenswrapper[4813]: I1206 15:57:00.689507 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-h5jv2" Dec 06 15:57:00 crc kubenswrapper[4813]: I1206 15:57:00.693316 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-f4x5w"] Dec 06 15:57:00 crc kubenswrapper[4813]: I1206 15:57:00.698133 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 06 15:57:00 crc kubenswrapper[4813]: I1206 15:57:00.706873 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-zw99x"] Dec 06 15:57:00 crc kubenswrapper[4813]: I1206 15:57:00.713023 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vc5f4"] Dec 06 15:57:00 crc kubenswrapper[4813]: I1206 15:57:00.713114 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-zw99x" Dec 06 15:57:00 crc kubenswrapper[4813]: I1206 15:57:00.795141 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/3b585cd5-6045-46b1-86cf-03536009e643-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-vc5f4\" (UID: \"3b585cd5-6045-46b1-86cf-03536009e643\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vc5f4" Dec 06 15:57:00 crc kubenswrapper[4813]: I1206 15:57:00.795217 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smswl\" (UniqueName: \"kubernetes.io/projected/db7dde13-c533-49ae-bae9-7239ad2b8db6-kube-api-access-smswl\") pod \"nmstate-metrics-7f946cbc9-f4x5w\" (UID: \"db7dde13-c533-49ae-bae9-7239ad2b8db6\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-f4x5w" Dec 06 15:57:00 crc kubenswrapper[4813]: I1206 15:57:00.795252 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7s5l2\" (UniqueName: \"kubernetes.io/projected/3b585cd5-6045-46b1-86cf-03536009e643-kube-api-access-7s5l2\") pod \"nmstate-webhook-5f6d4c5ccb-vc5f4\" (UID: \"3b585cd5-6045-46b1-86cf-03536009e643\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vc5f4" Dec 06 15:57:00 crc kubenswrapper[4813]: I1206 15:57:00.880094 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r7rqn"] Dec 06 15:57:00 crc kubenswrapper[4813]: I1206 15:57:00.880848 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r7rqn" Dec 06 15:57:00 crc kubenswrapper[4813]: W1206 15:57:00.888628 4813 reflector.go:561] object-"openshift-nmstate"/"nginx-conf": failed to list *v1.ConfigMap: configmaps "nginx-conf" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-nmstate": no relationship found between node 'crc' and this object Dec 06 15:57:00 crc kubenswrapper[4813]: E1206 15:57:00.888667 4813 reflector.go:158] "Unhandled Error" err="object-\"openshift-nmstate\"/\"nginx-conf\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"nginx-conf\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-nmstate\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 06 15:57:00 crc kubenswrapper[4813]: I1206 15:57:00.888722 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 06 15:57:00 crc kubenswrapper[4813]: W1206 15:57:00.891878 4813 reflector.go:561] object-"openshift-nmstate"/"default-dockercfg-9hn9w": failed to list *v1.Secret: secrets "default-dockercfg-9hn9w" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-nmstate": no relationship found between node 'crc' and this object Dec 06 15:57:00 crc kubenswrapper[4813]: E1206 15:57:00.891904 4813 reflector.go:158] "Unhandled Error" err="object-\"openshift-nmstate\"/\"default-dockercfg-9hn9w\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"default-dockercfg-9hn9w\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-nmstate\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 06 15:57:00 crc kubenswrapper[4813]: I1206 15:57:00.899407 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/3b585cd5-6045-46b1-86cf-03536009e643-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-vc5f4\" (UID: \"3b585cd5-6045-46b1-86cf-03536009e643\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vc5f4" Dec 06 15:57:00 crc kubenswrapper[4813]: I1206 15:57:00.899456 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/4f31e85e-1c68-4627-b047-72de6b05d490-nmstate-lock\") pod \"nmstate-handler-zw99x\" (UID: \"4f31e85e-1c68-4627-b047-72de6b05d490\") " pod="openshift-nmstate/nmstate-handler-zw99x" Dec 06 15:57:00 crc kubenswrapper[4813]: I1206 15:57:00.899477 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/4f31e85e-1c68-4627-b047-72de6b05d490-ovs-socket\") pod \"nmstate-handler-zw99x\" (UID: \"4f31e85e-1c68-4627-b047-72de6b05d490\") " pod="openshift-nmstate/nmstate-handler-zw99x" Dec 06 15:57:00 crc kubenswrapper[4813]: I1206 15:57:00.899519 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smswl\" (UniqueName: \"kubernetes.io/projected/db7dde13-c533-49ae-bae9-7239ad2b8db6-kube-api-access-smswl\") pod \"nmstate-metrics-7f946cbc9-f4x5w\" (UID: \"db7dde13-c533-49ae-bae9-7239ad2b8db6\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-f4x5w" Dec 06 15:57:00 crc kubenswrapper[4813]: I1206 15:57:00.899540 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bthb\" (UniqueName: \"kubernetes.io/projected/4f31e85e-1c68-4627-b047-72de6b05d490-kube-api-access-5bthb\") pod \"nmstate-handler-zw99x\" (UID: \"4f31e85e-1c68-4627-b047-72de6b05d490\") " pod="openshift-nmstate/nmstate-handler-zw99x" Dec 06 15:57:00 crc kubenswrapper[4813]: E1206 15:57:00.899559 4813 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 06 15:57:00 crc kubenswrapper[4813]: I1206 15:57:00.899574 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7s5l2\" (UniqueName: \"kubernetes.io/projected/3b585cd5-6045-46b1-86cf-03536009e643-kube-api-access-7s5l2\") pod \"nmstate-webhook-5f6d4c5ccb-vc5f4\" (UID: \"3b585cd5-6045-46b1-86cf-03536009e643\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vc5f4" Dec 06 15:57:00 crc kubenswrapper[4813]: E1206 15:57:00.899609 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3b585cd5-6045-46b1-86cf-03536009e643-tls-key-pair podName:3b585cd5-6045-46b1-86cf-03536009e643 nodeName:}" failed. No retries permitted until 2025-12-06 15:57:01.399594601 +0000 UTC m=+661.290474177 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/3b585cd5-6045-46b1-86cf-03536009e643-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-vc5f4" (UID: "3b585cd5-6045-46b1-86cf-03536009e643") : secret "openshift-nmstate-webhook" not found Dec 06 15:57:00 crc kubenswrapper[4813]: I1206 15:57:00.899633 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/4f31e85e-1c68-4627-b047-72de6b05d490-dbus-socket\") pod \"nmstate-handler-zw99x\" (UID: \"4f31e85e-1c68-4627-b047-72de6b05d490\") " pod="openshift-nmstate/nmstate-handler-zw99x" Dec 06 15:57:00 crc kubenswrapper[4813]: I1206 15:57:00.900414 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r7rqn"] Dec 06 15:57:00 crc kubenswrapper[4813]: I1206 15:57:00.929104 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smswl\" (UniqueName: \"kubernetes.io/projected/db7dde13-c533-49ae-bae9-7239ad2b8db6-kube-api-access-smswl\") pod \"nmstate-metrics-7f946cbc9-f4x5w\" (UID: \"db7dde13-c533-49ae-bae9-7239ad2b8db6\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-f4x5w" Dec 06 15:57:00 crc kubenswrapper[4813]: I1206 15:57:00.931684 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7s5l2\" (UniqueName: \"kubernetes.io/projected/3b585cd5-6045-46b1-86cf-03536009e643-kube-api-access-7s5l2\") pod \"nmstate-webhook-5f6d4c5ccb-vc5f4\" (UID: \"3b585cd5-6045-46b1-86cf-03536009e643\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vc5f4" Dec 06 15:57:00 crc kubenswrapper[4813]: I1206 15:57:00.979731 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-f4x5w" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.000251 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/4f31e85e-1c68-4627-b047-72de6b05d490-dbus-socket\") pod \"nmstate-handler-zw99x\" (UID: \"4f31e85e-1c68-4627-b047-72de6b05d490\") " pod="openshift-nmstate/nmstate-handler-zw99x" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.000329 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/4f31e85e-1c68-4627-b047-72de6b05d490-nmstate-lock\") pod \"nmstate-handler-zw99x\" (UID: \"4f31e85e-1c68-4627-b047-72de6b05d490\") " pod="openshift-nmstate/nmstate-handler-zw99x" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.000350 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/4f31e85e-1c68-4627-b047-72de6b05d490-ovs-socket\") pod \"nmstate-handler-zw99x\" (UID: \"4f31e85e-1c68-4627-b047-72de6b05d490\") " pod="openshift-nmstate/nmstate-handler-zw99x" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.000389 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bthb\" (UniqueName: \"kubernetes.io/projected/4f31e85e-1c68-4627-b047-72de6b05d490-kube-api-access-5bthb\") pod \"nmstate-handler-zw99x\" (UID: \"4f31e85e-1c68-4627-b047-72de6b05d490\") " pod="openshift-nmstate/nmstate-handler-zw99x" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.000407 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6w86\" (UniqueName: \"kubernetes.io/projected/2e477e7f-f00a-40a3-b26a-d49ef1f7f41d-kube-api-access-z6w86\") pod \"nmstate-console-plugin-7fbb5f6569-r7rqn\" (UID: \"2e477e7f-f00a-40a3-b26a-d49ef1f7f41d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r7rqn" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.000438 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/2e477e7f-f00a-40a3-b26a-d49ef1f7f41d-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-r7rqn\" (UID: \"2e477e7f-f00a-40a3-b26a-d49ef1f7f41d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r7rqn" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.000461 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/2e477e7f-f00a-40a3-b26a-d49ef1f7f41d-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-r7rqn\" (UID: \"2e477e7f-f00a-40a3-b26a-d49ef1f7f41d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r7rqn" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.000677 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/4f31e85e-1c68-4627-b047-72de6b05d490-dbus-socket\") pod \"nmstate-handler-zw99x\" (UID: \"4f31e85e-1c68-4627-b047-72de6b05d490\") " pod="openshift-nmstate/nmstate-handler-zw99x" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.000710 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/4f31e85e-1c68-4627-b047-72de6b05d490-nmstate-lock\") pod \"nmstate-handler-zw99x\" (UID: \"4f31e85e-1c68-4627-b047-72de6b05d490\") " pod="openshift-nmstate/nmstate-handler-zw99x" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.000730 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/4f31e85e-1c68-4627-b047-72de6b05d490-ovs-socket\") pod \"nmstate-handler-zw99x\" (UID: \"4f31e85e-1c68-4627-b047-72de6b05d490\") " pod="openshift-nmstate/nmstate-handler-zw99x" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.025873 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bthb\" (UniqueName: \"kubernetes.io/projected/4f31e85e-1c68-4627-b047-72de6b05d490-kube-api-access-5bthb\") pod \"nmstate-handler-zw99x\" (UID: \"4f31e85e-1c68-4627-b047-72de6b05d490\") " pod="openshift-nmstate/nmstate-handler-zw99x" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.032939 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-zw99x" Dec 06 15:57:01 crc kubenswrapper[4813]: W1206 15:57:01.056716 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4f31e85e_1c68_4627_b047_72de6b05d490.slice/crio-c4fc5706c204defc235feca1fab4adeb9f13f43a8dc5af8d3b5031a6b23d68a7 WatchSource:0}: Error finding container c4fc5706c204defc235feca1fab4adeb9f13f43a8dc5af8d3b5031a6b23d68a7: Status 404 returned error can't find the container with id c4fc5706c204defc235feca1fab4adeb9f13f43a8dc5af8d3b5031a6b23d68a7 Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.101380 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/2e477e7f-f00a-40a3-b26a-d49ef1f7f41d-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-r7rqn\" (UID: \"2e477e7f-f00a-40a3-b26a-d49ef1f7f41d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r7rqn" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.101427 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/2e477e7f-f00a-40a3-b26a-d49ef1f7f41d-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-r7rqn\" (UID: \"2e477e7f-f00a-40a3-b26a-d49ef1f7f41d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r7rqn" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.101501 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6w86\" (UniqueName: \"kubernetes.io/projected/2e477e7f-f00a-40a3-b26a-d49ef1f7f41d-kube-api-access-z6w86\") pod \"nmstate-console-plugin-7fbb5f6569-r7rqn\" (UID: \"2e477e7f-f00a-40a3-b26a-d49ef1f7f41d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r7rqn" Dec 06 15:57:01 crc kubenswrapper[4813]: E1206 15:57:01.101845 4813 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 06 15:57:01 crc kubenswrapper[4813]: E1206 15:57:01.101884 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2e477e7f-f00a-40a3-b26a-d49ef1f7f41d-plugin-serving-cert podName:2e477e7f-f00a-40a3-b26a-d49ef1f7f41d nodeName:}" failed. No retries permitted until 2025-12-06 15:57:01.601870964 +0000 UTC m=+661.492750540 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/2e477e7f-f00a-40a3-b26a-d49ef1f7f41d-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-r7rqn" (UID: "2e477e7f-f00a-40a3-b26a-d49ef1f7f41d") : secret "plugin-serving-cert" not found Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.115640 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7f86fb7b6-5hltw"] Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.118749 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7f86fb7b6-5hltw" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.140254 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6w86\" (UniqueName: \"kubernetes.io/projected/2e477e7f-f00a-40a3-b26a-d49ef1f7f41d-kube-api-access-z6w86\") pod \"nmstate-console-plugin-7fbb5f6569-r7rqn\" (UID: \"2e477e7f-f00a-40a3-b26a-d49ef1f7f41d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r7rqn" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.194542 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7f86fb7b6-5hltw"] Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.303845 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/71135043-dad1-4a59-a87f-650dda6af096-console-oauth-config\") pod \"console-7f86fb7b6-5hltw\" (UID: \"71135043-dad1-4a59-a87f-650dda6af096\") " pod="openshift-console/console-7f86fb7b6-5hltw" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.303900 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/71135043-dad1-4a59-a87f-650dda6af096-service-ca\") pod \"console-7f86fb7b6-5hltw\" (UID: \"71135043-dad1-4a59-a87f-650dda6af096\") " pod="openshift-console/console-7f86fb7b6-5hltw" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.303980 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgtsg\" (UniqueName: \"kubernetes.io/projected/71135043-dad1-4a59-a87f-650dda6af096-kube-api-access-cgtsg\") pod \"console-7f86fb7b6-5hltw\" (UID: \"71135043-dad1-4a59-a87f-650dda6af096\") " pod="openshift-console/console-7f86fb7b6-5hltw" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.304014 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/71135043-dad1-4a59-a87f-650dda6af096-oauth-serving-cert\") pod \"console-7f86fb7b6-5hltw\" (UID: \"71135043-dad1-4a59-a87f-650dda6af096\") " pod="openshift-console/console-7f86fb7b6-5hltw" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.304110 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/71135043-dad1-4a59-a87f-650dda6af096-trusted-ca-bundle\") pod \"console-7f86fb7b6-5hltw\" (UID: \"71135043-dad1-4a59-a87f-650dda6af096\") " pod="openshift-console/console-7f86fb7b6-5hltw" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.304140 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/71135043-dad1-4a59-a87f-650dda6af096-console-serving-cert\") pod \"console-7f86fb7b6-5hltw\" (UID: \"71135043-dad1-4a59-a87f-650dda6af096\") " pod="openshift-console/console-7f86fb7b6-5hltw" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.304156 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/71135043-dad1-4a59-a87f-650dda6af096-console-config\") pod \"console-7f86fb7b6-5hltw\" (UID: \"71135043-dad1-4a59-a87f-650dda6af096\") " pod="openshift-console/console-7f86fb7b6-5hltw" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.404950 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/71135043-dad1-4a59-a87f-650dda6af096-console-oauth-config\") pod \"console-7f86fb7b6-5hltw\" (UID: \"71135043-dad1-4a59-a87f-650dda6af096\") " pod="openshift-console/console-7f86fb7b6-5hltw" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.405059 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/71135043-dad1-4a59-a87f-650dda6af096-service-ca\") pod \"console-7f86fb7b6-5hltw\" (UID: \"71135043-dad1-4a59-a87f-650dda6af096\") " pod="openshift-console/console-7f86fb7b6-5hltw" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.405134 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgtsg\" (UniqueName: \"kubernetes.io/projected/71135043-dad1-4a59-a87f-650dda6af096-kube-api-access-cgtsg\") pod \"console-7f86fb7b6-5hltw\" (UID: \"71135043-dad1-4a59-a87f-650dda6af096\") " pod="openshift-console/console-7f86fb7b6-5hltw" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.405183 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/3b585cd5-6045-46b1-86cf-03536009e643-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-vc5f4\" (UID: \"3b585cd5-6045-46b1-86cf-03536009e643\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vc5f4" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.405222 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/71135043-dad1-4a59-a87f-650dda6af096-oauth-serving-cert\") pod \"console-7f86fb7b6-5hltw\" (UID: \"71135043-dad1-4a59-a87f-650dda6af096\") " pod="openshift-console/console-7f86fb7b6-5hltw" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.405352 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/71135043-dad1-4a59-a87f-650dda6af096-trusted-ca-bundle\") pod \"console-7f86fb7b6-5hltw\" (UID: \"71135043-dad1-4a59-a87f-650dda6af096\") " pod="openshift-console/console-7f86fb7b6-5hltw" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.405427 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/71135043-dad1-4a59-a87f-650dda6af096-console-serving-cert\") pod \"console-7f86fb7b6-5hltw\" (UID: \"71135043-dad1-4a59-a87f-650dda6af096\") " pod="openshift-console/console-7f86fb7b6-5hltw" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.405469 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/71135043-dad1-4a59-a87f-650dda6af096-console-config\") pod \"console-7f86fb7b6-5hltw\" (UID: \"71135043-dad1-4a59-a87f-650dda6af096\") " pod="openshift-console/console-7f86fb7b6-5hltw" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.406374 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/71135043-dad1-4a59-a87f-650dda6af096-oauth-serving-cert\") pod \"console-7f86fb7b6-5hltw\" (UID: \"71135043-dad1-4a59-a87f-650dda6af096\") " pod="openshift-console/console-7f86fb7b6-5hltw" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.406408 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/71135043-dad1-4a59-a87f-650dda6af096-service-ca\") pod \"console-7f86fb7b6-5hltw\" (UID: \"71135043-dad1-4a59-a87f-650dda6af096\") " pod="openshift-console/console-7f86fb7b6-5hltw" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.406847 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/71135043-dad1-4a59-a87f-650dda6af096-console-config\") pod \"console-7f86fb7b6-5hltw\" (UID: \"71135043-dad1-4a59-a87f-650dda6af096\") " pod="openshift-console/console-7f86fb7b6-5hltw" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.406916 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/71135043-dad1-4a59-a87f-650dda6af096-trusted-ca-bundle\") pod \"console-7f86fb7b6-5hltw\" (UID: \"71135043-dad1-4a59-a87f-650dda6af096\") " pod="openshift-console/console-7f86fb7b6-5hltw" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.408575 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/71135043-dad1-4a59-a87f-650dda6af096-console-oauth-config\") pod \"console-7f86fb7b6-5hltw\" (UID: \"71135043-dad1-4a59-a87f-650dda6af096\") " pod="openshift-console/console-7f86fb7b6-5hltw" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.409747 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/71135043-dad1-4a59-a87f-650dda6af096-console-serving-cert\") pod \"console-7f86fb7b6-5hltw\" (UID: \"71135043-dad1-4a59-a87f-650dda6af096\") " pod="openshift-console/console-7f86fb7b6-5hltw" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.412060 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/3b585cd5-6045-46b1-86cf-03536009e643-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-vc5f4\" (UID: \"3b585cd5-6045-46b1-86cf-03536009e643\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vc5f4" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.422401 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgtsg\" (UniqueName: \"kubernetes.io/projected/71135043-dad1-4a59-a87f-650dda6af096-kube-api-access-cgtsg\") pod \"console-7f86fb7b6-5hltw\" (UID: \"71135043-dad1-4a59-a87f-650dda6af096\") " pod="openshift-console/console-7f86fb7b6-5hltw" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.443177 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7f86fb7b6-5hltw" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.478829 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-f4x5w"] Dec 06 15:57:01 crc kubenswrapper[4813]: W1206 15:57:01.487233 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb7dde13_c533_49ae_bae9_7239ad2b8db6.slice/crio-92e4b2e1b0b287ba42e6300c2bed8ed1222b911722312477b0a8c0b65b7b386c WatchSource:0}: Error finding container 92e4b2e1b0b287ba42e6300c2bed8ed1222b911722312477b0a8c0b65b7b386c: Status 404 returned error can't find the container with id 92e4b2e1b0b287ba42e6300c2bed8ed1222b911722312477b0a8c0b65b7b386c Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.607877 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/2e477e7f-f00a-40a3-b26a-d49ef1f7f41d-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-r7rqn\" (UID: \"2e477e7f-f00a-40a3-b26a-d49ef1f7f41d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r7rqn" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.611945 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/2e477e7f-f00a-40a3-b26a-d49ef1f7f41d-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-r7rqn\" (UID: \"2e477e7f-f00a-40a3-b26a-d49ef1f7f41d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r7rqn" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.613918 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vc5f4" Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.691251 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-zw99x" event={"ID":"4f31e85e-1c68-4627-b047-72de6b05d490","Type":"ContainerStarted","Data":"c4fc5706c204defc235feca1fab4adeb9f13f43a8dc5af8d3b5031a6b23d68a7"} Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.691887 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-f4x5w" event={"ID":"db7dde13-c533-49ae-bae9-7239ad2b8db6","Type":"ContainerStarted","Data":"92e4b2e1b0b287ba42e6300c2bed8ed1222b911722312477b0a8c0b65b7b386c"} Dec 06 15:57:01 crc kubenswrapper[4813]: I1206 15:57:01.852419 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7f86fb7b6-5hltw"] Dec 06 15:57:01 crc kubenswrapper[4813]: W1206 15:57:01.855079 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71135043_dad1_4a59_a87f_650dda6af096.slice/crio-378fac93570881dddba086fb8fb004b6d8e055b3874378843f0ee8f4e6b10e39 WatchSource:0}: Error finding container 378fac93570881dddba086fb8fb004b6d8e055b3874378843f0ee8f4e6b10e39: Status 404 returned error can't find the container with id 378fac93570881dddba086fb8fb004b6d8e055b3874378843f0ee8f4e6b10e39 Dec 06 15:57:02 crc kubenswrapper[4813]: I1206 15:57:02.053034 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vc5f4"] Dec 06 15:57:02 crc kubenswrapper[4813]: E1206 15:57:02.103055 4813 configmap.go:193] Couldn't get configMap openshift-nmstate/nginx-conf: failed to sync configmap cache: timed out waiting for the condition Dec 06 15:57:02 crc kubenswrapper[4813]: E1206 15:57:02.103134 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2e477e7f-f00a-40a3-b26a-d49ef1f7f41d-nginx-conf podName:2e477e7f-f00a-40a3-b26a-d49ef1f7f41d nodeName:}" failed. No retries permitted until 2025-12-06 15:57:02.603116023 +0000 UTC m=+662.493995599 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/2e477e7f-f00a-40a3-b26a-d49ef1f7f41d-nginx-conf") pod "nmstate-console-plugin-7fbb5f6569-r7rqn" (UID: "2e477e7f-f00a-40a3-b26a-d49ef1f7f41d") : failed to sync configmap cache: timed out waiting for the condition Dec 06 15:57:02 crc kubenswrapper[4813]: I1206 15:57:02.364810 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-9hn9w" Dec 06 15:57:02 crc kubenswrapper[4813]: I1206 15:57:02.422512 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 06 15:57:02 crc kubenswrapper[4813]: I1206 15:57:02.620239 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/2e477e7f-f00a-40a3-b26a-d49ef1f7f41d-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-r7rqn\" (UID: \"2e477e7f-f00a-40a3-b26a-d49ef1f7f41d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r7rqn" Dec 06 15:57:02 crc kubenswrapper[4813]: I1206 15:57:02.621029 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/2e477e7f-f00a-40a3-b26a-d49ef1f7f41d-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-r7rqn\" (UID: \"2e477e7f-f00a-40a3-b26a-d49ef1f7f41d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r7rqn" Dec 06 15:57:02 crc kubenswrapper[4813]: I1206 15:57:02.699061 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r7rqn" Dec 06 15:57:02 crc kubenswrapper[4813]: I1206 15:57:02.699065 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vc5f4" event={"ID":"3b585cd5-6045-46b1-86cf-03536009e643","Type":"ContainerStarted","Data":"e334890edfdd95d98396556d9b7651622258eb2a9133edd29a25d4bd8e00a9d8"} Dec 06 15:57:02 crc kubenswrapper[4813]: I1206 15:57:02.701584 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7f86fb7b6-5hltw" event={"ID":"71135043-dad1-4a59-a87f-650dda6af096","Type":"ContainerStarted","Data":"f6ea355550c077c66304f974a8c48e02e0a107f43eade33dcc8f408bc80c5bd2"} Dec 06 15:57:02 crc kubenswrapper[4813]: I1206 15:57:02.701618 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7f86fb7b6-5hltw" event={"ID":"71135043-dad1-4a59-a87f-650dda6af096","Type":"ContainerStarted","Data":"378fac93570881dddba086fb8fb004b6d8e055b3874378843f0ee8f4e6b10e39"} Dec 06 15:57:02 crc kubenswrapper[4813]: I1206 15:57:02.719374 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7f86fb7b6-5hltw" podStartSLOduration=1.719359169 podStartE2EDuration="1.719359169s" podCreationTimestamp="2025-12-06 15:57:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:57:02.717390257 +0000 UTC m=+662.608269833" watchObservedRunningTime="2025-12-06 15:57:02.719359169 +0000 UTC m=+662.610238745" Dec 06 15:57:03 crc kubenswrapper[4813]: I1206 15:57:03.501284 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r7rqn"] Dec 06 15:57:03 crc kubenswrapper[4813]: I1206 15:57:03.715752 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-zw99x" event={"ID":"4f31e85e-1c68-4627-b047-72de6b05d490","Type":"ContainerStarted","Data":"3b99a154588f0cbf41d1b972096e3ad307585d37ec90a7e2ef4fe901f701bc39"} Dec 06 15:57:03 crc kubenswrapper[4813]: I1206 15:57:03.715891 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-zw99x" Dec 06 15:57:03 crc kubenswrapper[4813]: I1206 15:57:03.717339 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-f4x5w" event={"ID":"db7dde13-c533-49ae-bae9-7239ad2b8db6","Type":"ContainerStarted","Data":"9d22d83e7bfa02b3fc033b156f8c96fca38c902db2e54dc3b1ded8064e819f08"} Dec 06 15:57:03 crc kubenswrapper[4813]: I1206 15:57:03.721470 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vc5f4" event={"ID":"3b585cd5-6045-46b1-86cf-03536009e643","Type":"ContainerStarted","Data":"b120962bd17cf82c42b036c2ad96aee617c070599d839091b4d08afea6a3d326"} Dec 06 15:57:03 crc kubenswrapper[4813]: I1206 15:57:03.721641 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vc5f4" Dec 06 15:57:03 crc kubenswrapper[4813]: I1206 15:57:03.722800 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r7rqn" event={"ID":"2e477e7f-f00a-40a3-b26a-d49ef1f7f41d","Type":"ContainerStarted","Data":"cd61759c1b429201333e87c422ac53786bc9c8e29fb6e54c90a1cb29a88559ba"} Dec 06 15:57:03 crc kubenswrapper[4813]: I1206 15:57:03.750712 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-zw99x" podStartSLOduration=1.583105534 podStartE2EDuration="3.750687088s" podCreationTimestamp="2025-12-06 15:57:00 +0000 UTC" firstStartedPulling="2025-12-06 15:57:01.059007998 +0000 UTC m=+660.949887574" lastFinishedPulling="2025-12-06 15:57:03.226589552 +0000 UTC m=+663.117469128" observedRunningTime="2025-12-06 15:57:03.738660672 +0000 UTC m=+663.629540278" watchObservedRunningTime="2025-12-06 15:57:03.750687088 +0000 UTC m=+663.641566704" Dec 06 15:57:03 crc kubenswrapper[4813]: I1206 15:57:03.768058 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vc5f4" podStartSLOduration=2.602679015 podStartE2EDuration="3.768031624s" podCreationTimestamp="2025-12-06 15:57:00 +0000 UTC" firstStartedPulling="2025-12-06 15:57:02.062211429 +0000 UTC m=+661.953091005" lastFinishedPulling="2025-12-06 15:57:03.227564038 +0000 UTC m=+663.118443614" observedRunningTime="2025-12-06 15:57:03.765495977 +0000 UTC m=+663.656375553" watchObservedRunningTime="2025-12-06 15:57:03.768031624 +0000 UTC m=+663.658911240" Dec 06 15:57:05 crc kubenswrapper[4813]: I1206 15:57:05.737768 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-f4x5w" event={"ID":"db7dde13-c533-49ae-bae9-7239ad2b8db6","Type":"ContainerStarted","Data":"536f7239a0eff4c9362e90cc1b40716c27811b89823803e3f191243d33dff76a"} Dec 06 15:57:05 crc kubenswrapper[4813]: I1206 15:57:05.741190 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r7rqn" event={"ID":"2e477e7f-f00a-40a3-b26a-d49ef1f7f41d","Type":"ContainerStarted","Data":"c26d0fac7c06a960a4f6da74f4229c13d4012ccbcde1eb69d3b3dc96a354cb7d"} Dec 06 15:57:05 crc kubenswrapper[4813]: I1206 15:57:05.757835 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-f4x5w" podStartSLOduration=1.732745815 podStartE2EDuration="5.757797978s" podCreationTimestamp="2025-12-06 15:57:00 +0000 UTC" firstStartedPulling="2025-12-06 15:57:01.490675067 +0000 UTC m=+661.381554683" lastFinishedPulling="2025-12-06 15:57:05.51572727 +0000 UTC m=+665.406606846" observedRunningTime="2025-12-06 15:57:05.753103835 +0000 UTC m=+665.643983411" watchObservedRunningTime="2025-12-06 15:57:05.757797978 +0000 UTC m=+665.648677554" Dec 06 15:57:05 crc kubenswrapper[4813]: I1206 15:57:05.840987 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r7rqn" podStartSLOduration=3.869543864 podStartE2EDuration="5.840956676s" podCreationTimestamp="2025-12-06 15:57:00 +0000 UTC" firstStartedPulling="2025-12-06 15:57:03.512006189 +0000 UTC m=+663.402885765" lastFinishedPulling="2025-12-06 15:57:05.483419001 +0000 UTC m=+665.374298577" observedRunningTime="2025-12-06 15:57:05.839121648 +0000 UTC m=+665.730001224" watchObservedRunningTime="2025-12-06 15:57:05.840956676 +0000 UTC m=+665.731836252" Dec 06 15:57:11 crc kubenswrapper[4813]: I1206 15:57:11.068424 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-zw99x" Dec 06 15:57:11 crc kubenswrapper[4813]: I1206 15:57:11.445391 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7f86fb7b6-5hltw" Dec 06 15:57:11 crc kubenswrapper[4813]: I1206 15:57:11.445885 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7f86fb7b6-5hltw" Dec 06 15:57:11 crc kubenswrapper[4813]: I1206 15:57:11.451085 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7f86fb7b6-5hltw" Dec 06 15:57:11 crc kubenswrapper[4813]: I1206 15:57:11.810210 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7f86fb7b6-5hltw" Dec 06 15:57:11 crc kubenswrapper[4813]: I1206 15:57:11.877892 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-gb6cc"] Dec 06 15:57:21 crc kubenswrapper[4813]: I1206 15:57:21.624397 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-vc5f4" Dec 06 15:57:36 crc kubenswrapper[4813]: I1206 15:57:36.325412 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5"] Dec 06 15:57:36 crc kubenswrapper[4813]: I1206 15:57:36.327574 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5" Dec 06 15:57:36 crc kubenswrapper[4813]: I1206 15:57:36.330863 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 06 15:57:36 crc kubenswrapper[4813]: I1206 15:57:36.343540 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5"] Dec 06 15:57:36 crc kubenswrapper[4813]: I1206 15:57:36.484351 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7b4p4\" (UniqueName: \"kubernetes.io/projected/6f21e1e0-2a67-4f45-93f4-e84a5bba3007-kube-api-access-7b4p4\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5\" (UID: \"6f21e1e0-2a67-4f45-93f4-e84a5bba3007\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5" Dec 06 15:57:36 crc kubenswrapper[4813]: I1206 15:57:36.484446 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6f21e1e0-2a67-4f45-93f4-e84a5bba3007-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5\" (UID: \"6f21e1e0-2a67-4f45-93f4-e84a5bba3007\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5" Dec 06 15:57:36 crc kubenswrapper[4813]: I1206 15:57:36.484575 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6f21e1e0-2a67-4f45-93f4-e84a5bba3007-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5\" (UID: \"6f21e1e0-2a67-4f45-93f4-e84a5bba3007\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5" Dec 06 15:57:36 crc kubenswrapper[4813]: I1206 15:57:36.586377 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6f21e1e0-2a67-4f45-93f4-e84a5bba3007-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5\" (UID: \"6f21e1e0-2a67-4f45-93f4-e84a5bba3007\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5" Dec 06 15:57:36 crc kubenswrapper[4813]: I1206 15:57:36.586474 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7b4p4\" (UniqueName: \"kubernetes.io/projected/6f21e1e0-2a67-4f45-93f4-e84a5bba3007-kube-api-access-7b4p4\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5\" (UID: \"6f21e1e0-2a67-4f45-93f4-e84a5bba3007\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5" Dec 06 15:57:36 crc kubenswrapper[4813]: I1206 15:57:36.586530 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6f21e1e0-2a67-4f45-93f4-e84a5bba3007-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5\" (UID: \"6f21e1e0-2a67-4f45-93f4-e84a5bba3007\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5" Dec 06 15:57:36 crc kubenswrapper[4813]: I1206 15:57:36.587250 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6f21e1e0-2a67-4f45-93f4-e84a5bba3007-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5\" (UID: \"6f21e1e0-2a67-4f45-93f4-e84a5bba3007\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5" Dec 06 15:57:36 crc kubenswrapper[4813]: I1206 15:57:36.587752 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6f21e1e0-2a67-4f45-93f4-e84a5bba3007-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5\" (UID: \"6f21e1e0-2a67-4f45-93f4-e84a5bba3007\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5" Dec 06 15:57:36 crc kubenswrapper[4813]: I1206 15:57:36.621941 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7b4p4\" (UniqueName: \"kubernetes.io/projected/6f21e1e0-2a67-4f45-93f4-e84a5bba3007-kube-api-access-7b4p4\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5\" (UID: \"6f21e1e0-2a67-4f45-93f4-e84a5bba3007\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5" Dec 06 15:57:36 crc kubenswrapper[4813]: I1206 15:57:36.648108 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5" Dec 06 15:57:36 crc kubenswrapper[4813]: I1206 15:57:36.944196 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-gb6cc" podUID="36f25c5a-3e57-4367-b306-db9661e4f7c9" containerName="console" containerID="cri-o://4861be593d3b9a7a02282408617c25a42bd7fd31cdca7fb4fd51a31d05090c08" gracePeriod=15 Dec 06 15:57:36 crc kubenswrapper[4813]: I1206 15:57:36.972545 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5"] Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.336970 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-gb6cc_36f25c5a-3e57-4367-b306-db9661e4f7c9/console/0.log" Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.337486 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-gb6cc" Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.502429 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/36f25c5a-3e57-4367-b306-db9661e4f7c9-oauth-serving-cert\") pod \"36f25c5a-3e57-4367-b306-db9661e4f7c9\" (UID: \"36f25c5a-3e57-4367-b306-db9661e4f7c9\") " Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.502480 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cncsp\" (UniqueName: \"kubernetes.io/projected/36f25c5a-3e57-4367-b306-db9661e4f7c9-kube-api-access-cncsp\") pod \"36f25c5a-3e57-4367-b306-db9661e4f7c9\" (UID: \"36f25c5a-3e57-4367-b306-db9661e4f7c9\") " Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.502530 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/36f25c5a-3e57-4367-b306-db9661e4f7c9-trusted-ca-bundle\") pod \"36f25c5a-3e57-4367-b306-db9661e4f7c9\" (UID: \"36f25c5a-3e57-4367-b306-db9661e4f7c9\") " Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.502558 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/36f25c5a-3e57-4367-b306-db9661e4f7c9-service-ca\") pod \"36f25c5a-3e57-4367-b306-db9661e4f7c9\" (UID: \"36f25c5a-3e57-4367-b306-db9661e4f7c9\") " Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.502598 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/36f25c5a-3e57-4367-b306-db9661e4f7c9-console-oauth-config\") pod \"36f25c5a-3e57-4367-b306-db9661e4f7c9\" (UID: \"36f25c5a-3e57-4367-b306-db9661e4f7c9\") " Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.502617 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/36f25c5a-3e57-4367-b306-db9661e4f7c9-console-config\") pod \"36f25c5a-3e57-4367-b306-db9661e4f7c9\" (UID: \"36f25c5a-3e57-4367-b306-db9661e4f7c9\") " Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.502658 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/36f25c5a-3e57-4367-b306-db9661e4f7c9-console-serving-cert\") pod \"36f25c5a-3e57-4367-b306-db9661e4f7c9\" (UID: \"36f25c5a-3e57-4367-b306-db9661e4f7c9\") " Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.503229 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36f25c5a-3e57-4367-b306-db9661e4f7c9-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "36f25c5a-3e57-4367-b306-db9661e4f7c9" (UID: "36f25c5a-3e57-4367-b306-db9661e4f7c9"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.503242 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36f25c5a-3e57-4367-b306-db9661e4f7c9-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "36f25c5a-3e57-4367-b306-db9661e4f7c9" (UID: "36f25c5a-3e57-4367-b306-db9661e4f7c9"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.503617 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36f25c5a-3e57-4367-b306-db9661e4f7c9-console-config" (OuterVolumeSpecName: "console-config") pod "36f25c5a-3e57-4367-b306-db9661e4f7c9" (UID: "36f25c5a-3e57-4367-b306-db9661e4f7c9"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.503667 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36f25c5a-3e57-4367-b306-db9661e4f7c9-service-ca" (OuterVolumeSpecName: "service-ca") pod "36f25c5a-3e57-4367-b306-db9661e4f7c9" (UID: "36f25c5a-3e57-4367-b306-db9661e4f7c9"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.508073 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36f25c5a-3e57-4367-b306-db9661e4f7c9-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "36f25c5a-3e57-4367-b306-db9661e4f7c9" (UID: "36f25c5a-3e57-4367-b306-db9661e4f7c9"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.508193 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36f25c5a-3e57-4367-b306-db9661e4f7c9-kube-api-access-cncsp" (OuterVolumeSpecName: "kube-api-access-cncsp") pod "36f25c5a-3e57-4367-b306-db9661e4f7c9" (UID: "36f25c5a-3e57-4367-b306-db9661e4f7c9"). InnerVolumeSpecName "kube-api-access-cncsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.508848 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36f25c5a-3e57-4367-b306-db9661e4f7c9-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "36f25c5a-3e57-4367-b306-db9661e4f7c9" (UID: "36f25c5a-3e57-4367-b306-db9661e4f7c9"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.604453 4813 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/36f25c5a-3e57-4367-b306-db9661e4f7c9-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.604495 4813 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/36f25c5a-3e57-4367-b306-db9661e4f7c9-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.604515 4813 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/36f25c5a-3e57-4367-b306-db9661e4f7c9-console-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.604533 4813 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/36f25c5a-3e57-4367-b306-db9661e4f7c9-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.604549 4813 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/36f25c5a-3e57-4367-b306-db9661e4f7c9-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.604564 4813 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/36f25c5a-3e57-4367-b306-db9661e4f7c9-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.604580 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cncsp\" (UniqueName: \"kubernetes.io/projected/36f25c5a-3e57-4367-b306-db9661e4f7c9-kube-api-access-cncsp\") on node \"crc\" DevicePath \"\"" Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.982976 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-gb6cc_36f25c5a-3e57-4367-b306-db9661e4f7c9/console/0.log" Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.983049 4813 generic.go:334] "Generic (PLEG): container finished" podID="36f25c5a-3e57-4367-b306-db9661e4f7c9" containerID="4861be593d3b9a7a02282408617c25a42bd7fd31cdca7fb4fd51a31d05090c08" exitCode=2 Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.983137 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-gb6cc" event={"ID":"36f25c5a-3e57-4367-b306-db9661e4f7c9","Type":"ContainerDied","Data":"4861be593d3b9a7a02282408617c25a42bd7fd31cdca7fb4fd51a31d05090c08"} Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.983174 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-gb6cc" event={"ID":"36f25c5a-3e57-4367-b306-db9661e4f7c9","Type":"ContainerDied","Data":"299d968a296378c9098c081177ae4a317e4111ddf0168a491ed8eefde4e399e6"} Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.983143 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-gb6cc" Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.983203 4813 scope.go:117] "RemoveContainer" containerID="4861be593d3b9a7a02282408617c25a42bd7fd31cdca7fb4fd51a31d05090c08" Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.986013 4813 generic.go:334] "Generic (PLEG): container finished" podID="6f21e1e0-2a67-4f45-93f4-e84a5bba3007" containerID="e1672cc1359bdcc4e07ce212bd1fe9402549eafe7df32de4659d4d8467d3ba06" exitCode=0 Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.986159 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5" event={"ID":"6f21e1e0-2a67-4f45-93f4-e84a5bba3007","Type":"ContainerDied","Data":"e1672cc1359bdcc4e07ce212bd1fe9402549eafe7df32de4659d4d8467d3ba06"} Dec 06 15:57:37 crc kubenswrapper[4813]: I1206 15:57:37.986198 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5" event={"ID":"6f21e1e0-2a67-4f45-93f4-e84a5bba3007","Type":"ContainerStarted","Data":"71c01e48d6002a2d5a88d2daead1289f0bf0b6de433359a0146231c3f1ee304b"} Dec 06 15:57:38 crc kubenswrapper[4813]: I1206 15:57:38.020194 4813 scope.go:117] "RemoveContainer" containerID="4861be593d3b9a7a02282408617c25a42bd7fd31cdca7fb4fd51a31d05090c08" Dec 06 15:57:38 crc kubenswrapper[4813]: E1206 15:57:38.021411 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4861be593d3b9a7a02282408617c25a42bd7fd31cdca7fb4fd51a31d05090c08\": container with ID starting with 4861be593d3b9a7a02282408617c25a42bd7fd31cdca7fb4fd51a31d05090c08 not found: ID does not exist" containerID="4861be593d3b9a7a02282408617c25a42bd7fd31cdca7fb4fd51a31d05090c08" Dec 06 15:57:38 crc kubenswrapper[4813]: I1206 15:57:38.022101 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4861be593d3b9a7a02282408617c25a42bd7fd31cdca7fb4fd51a31d05090c08"} err="failed to get container status \"4861be593d3b9a7a02282408617c25a42bd7fd31cdca7fb4fd51a31d05090c08\": rpc error: code = NotFound desc = could not find container \"4861be593d3b9a7a02282408617c25a42bd7fd31cdca7fb4fd51a31d05090c08\": container with ID starting with 4861be593d3b9a7a02282408617c25a42bd7fd31cdca7fb4fd51a31d05090c08 not found: ID does not exist" Dec 06 15:57:38 crc kubenswrapper[4813]: I1206 15:57:38.047384 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-gb6cc"] Dec 06 15:57:38 crc kubenswrapper[4813]: I1206 15:57:38.053216 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-gb6cc"] Dec 06 15:57:38 crc kubenswrapper[4813]: I1206 15:57:38.502677 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36f25c5a-3e57-4367-b306-db9661e4f7c9" path="/var/lib/kubelet/pods/36f25c5a-3e57-4367-b306-db9661e4f7c9/volumes" Dec 06 15:57:40 crc kubenswrapper[4813]: I1206 15:57:40.007725 4813 generic.go:334] "Generic (PLEG): container finished" podID="6f21e1e0-2a67-4f45-93f4-e84a5bba3007" containerID="1ebb8090c54584c68731557edb837fad2955f2798f53dbd75db416721f918ffe" exitCode=0 Dec 06 15:57:40 crc kubenswrapper[4813]: I1206 15:57:40.008033 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5" event={"ID":"6f21e1e0-2a67-4f45-93f4-e84a5bba3007","Type":"ContainerDied","Data":"1ebb8090c54584c68731557edb837fad2955f2798f53dbd75db416721f918ffe"} Dec 06 15:57:41 crc kubenswrapper[4813]: I1206 15:57:41.020625 4813 generic.go:334] "Generic (PLEG): container finished" podID="6f21e1e0-2a67-4f45-93f4-e84a5bba3007" containerID="a7c4d330091a91104c97fe71869492bbbcf9decd3de4dea864af0ad9f5cefaf1" exitCode=0 Dec 06 15:57:41 crc kubenswrapper[4813]: I1206 15:57:41.020706 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5" event={"ID":"6f21e1e0-2a67-4f45-93f4-e84a5bba3007","Type":"ContainerDied","Data":"a7c4d330091a91104c97fe71869492bbbcf9decd3de4dea864af0ad9f5cefaf1"} Dec 06 15:57:42 crc kubenswrapper[4813]: I1206 15:57:42.348225 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5" Dec 06 15:57:42 crc kubenswrapper[4813]: I1206 15:57:42.472415 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6f21e1e0-2a67-4f45-93f4-e84a5bba3007-bundle\") pod \"6f21e1e0-2a67-4f45-93f4-e84a5bba3007\" (UID: \"6f21e1e0-2a67-4f45-93f4-e84a5bba3007\") " Dec 06 15:57:42 crc kubenswrapper[4813]: I1206 15:57:42.472628 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6f21e1e0-2a67-4f45-93f4-e84a5bba3007-util\") pod \"6f21e1e0-2a67-4f45-93f4-e84a5bba3007\" (UID: \"6f21e1e0-2a67-4f45-93f4-e84a5bba3007\") " Dec 06 15:57:42 crc kubenswrapper[4813]: I1206 15:57:42.472734 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7b4p4\" (UniqueName: \"kubernetes.io/projected/6f21e1e0-2a67-4f45-93f4-e84a5bba3007-kube-api-access-7b4p4\") pod \"6f21e1e0-2a67-4f45-93f4-e84a5bba3007\" (UID: \"6f21e1e0-2a67-4f45-93f4-e84a5bba3007\") " Dec 06 15:57:42 crc kubenswrapper[4813]: I1206 15:57:42.473742 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f21e1e0-2a67-4f45-93f4-e84a5bba3007-bundle" (OuterVolumeSpecName: "bundle") pod "6f21e1e0-2a67-4f45-93f4-e84a5bba3007" (UID: "6f21e1e0-2a67-4f45-93f4-e84a5bba3007"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:57:42 crc kubenswrapper[4813]: I1206 15:57:42.481080 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f21e1e0-2a67-4f45-93f4-e84a5bba3007-kube-api-access-7b4p4" (OuterVolumeSpecName: "kube-api-access-7b4p4") pod "6f21e1e0-2a67-4f45-93f4-e84a5bba3007" (UID: "6f21e1e0-2a67-4f45-93f4-e84a5bba3007"). InnerVolumeSpecName "kube-api-access-7b4p4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:57:42 crc kubenswrapper[4813]: I1206 15:57:42.574914 4813 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6f21e1e0-2a67-4f45-93f4-e84a5bba3007-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 15:57:42 crc kubenswrapper[4813]: I1206 15:57:42.574953 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7b4p4\" (UniqueName: \"kubernetes.io/projected/6f21e1e0-2a67-4f45-93f4-e84a5bba3007-kube-api-access-7b4p4\") on node \"crc\" DevicePath \"\"" Dec 06 15:57:42 crc kubenswrapper[4813]: I1206 15:57:42.584190 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f21e1e0-2a67-4f45-93f4-e84a5bba3007-util" (OuterVolumeSpecName: "util") pod "6f21e1e0-2a67-4f45-93f4-e84a5bba3007" (UID: "6f21e1e0-2a67-4f45-93f4-e84a5bba3007"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:57:42 crc kubenswrapper[4813]: I1206 15:57:42.676443 4813 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6f21e1e0-2a67-4f45-93f4-e84a5bba3007-util\") on node \"crc\" DevicePath \"\"" Dec 06 15:57:43 crc kubenswrapper[4813]: I1206 15:57:43.039762 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5" event={"ID":"6f21e1e0-2a67-4f45-93f4-e84a5bba3007","Type":"ContainerDied","Data":"71c01e48d6002a2d5a88d2daead1289f0bf0b6de433359a0146231c3f1ee304b"} Dec 06 15:57:43 crc kubenswrapper[4813]: I1206 15:57:43.039806 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71c01e48d6002a2d5a88d2daead1289f0bf0b6de433359a0146231c3f1ee304b" Dec 06 15:57:43 crc kubenswrapper[4813]: I1206 15:57:43.039824 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.391900 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-6977db5dfd-44msc"] Dec 06 15:57:51 crc kubenswrapper[4813]: E1206 15:57:51.392678 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f21e1e0-2a67-4f45-93f4-e84a5bba3007" containerName="pull" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.392696 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f21e1e0-2a67-4f45-93f4-e84a5bba3007" containerName="pull" Dec 06 15:57:51 crc kubenswrapper[4813]: E1206 15:57:51.392717 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36f25c5a-3e57-4367-b306-db9661e4f7c9" containerName="console" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.392726 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="36f25c5a-3e57-4367-b306-db9661e4f7c9" containerName="console" Dec 06 15:57:51 crc kubenswrapper[4813]: E1206 15:57:51.392737 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f21e1e0-2a67-4f45-93f4-e84a5bba3007" containerName="util" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.392747 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f21e1e0-2a67-4f45-93f4-e84a5bba3007" containerName="util" Dec 06 15:57:51 crc kubenswrapper[4813]: E1206 15:57:51.392763 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f21e1e0-2a67-4f45-93f4-e84a5bba3007" containerName="extract" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.392770 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f21e1e0-2a67-4f45-93f4-e84a5bba3007" containerName="extract" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.392889 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f21e1e0-2a67-4f45-93f4-e84a5bba3007" containerName="extract" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.392902 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="36f25c5a-3e57-4367-b306-db9661e4f7c9" containerName="console" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.393365 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6977db5dfd-44msc" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.397469 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.397581 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.397621 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.397695 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-zfl42" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.397770 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.465134 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6977db5dfd-44msc"] Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.484964 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e7537632-d133-4c59-a167-e59127ef6fab-webhook-cert\") pod \"metallb-operator-controller-manager-6977db5dfd-44msc\" (UID: \"e7537632-d133-4c59-a167-e59127ef6fab\") " pod="metallb-system/metallb-operator-controller-manager-6977db5dfd-44msc" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.485031 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e7537632-d133-4c59-a167-e59127ef6fab-apiservice-cert\") pod \"metallb-operator-controller-manager-6977db5dfd-44msc\" (UID: \"e7537632-d133-4c59-a167-e59127ef6fab\") " pod="metallb-system/metallb-operator-controller-manager-6977db5dfd-44msc" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.485214 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjbkt\" (UniqueName: \"kubernetes.io/projected/e7537632-d133-4c59-a167-e59127ef6fab-kube-api-access-pjbkt\") pod \"metallb-operator-controller-manager-6977db5dfd-44msc\" (UID: \"e7537632-d133-4c59-a167-e59127ef6fab\") " pod="metallb-system/metallb-operator-controller-manager-6977db5dfd-44msc" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.586044 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e7537632-d133-4c59-a167-e59127ef6fab-webhook-cert\") pod \"metallb-operator-controller-manager-6977db5dfd-44msc\" (UID: \"e7537632-d133-4c59-a167-e59127ef6fab\") " pod="metallb-system/metallb-operator-controller-manager-6977db5dfd-44msc" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.586118 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e7537632-d133-4c59-a167-e59127ef6fab-apiservice-cert\") pod \"metallb-operator-controller-manager-6977db5dfd-44msc\" (UID: \"e7537632-d133-4c59-a167-e59127ef6fab\") " pod="metallb-system/metallb-operator-controller-manager-6977db5dfd-44msc" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.586171 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjbkt\" (UniqueName: \"kubernetes.io/projected/e7537632-d133-4c59-a167-e59127ef6fab-kube-api-access-pjbkt\") pod \"metallb-operator-controller-manager-6977db5dfd-44msc\" (UID: \"e7537632-d133-4c59-a167-e59127ef6fab\") " pod="metallb-system/metallb-operator-controller-manager-6977db5dfd-44msc" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.592291 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e7537632-d133-4c59-a167-e59127ef6fab-webhook-cert\") pod \"metallb-operator-controller-manager-6977db5dfd-44msc\" (UID: \"e7537632-d133-4c59-a167-e59127ef6fab\") " pod="metallb-system/metallb-operator-controller-manager-6977db5dfd-44msc" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.592312 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e7537632-d133-4c59-a167-e59127ef6fab-apiservice-cert\") pod \"metallb-operator-controller-manager-6977db5dfd-44msc\" (UID: \"e7537632-d133-4c59-a167-e59127ef6fab\") " pod="metallb-system/metallb-operator-controller-manager-6977db5dfd-44msc" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.604324 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjbkt\" (UniqueName: \"kubernetes.io/projected/e7537632-d133-4c59-a167-e59127ef6fab-kube-api-access-pjbkt\") pod \"metallb-operator-controller-manager-6977db5dfd-44msc\" (UID: \"e7537632-d133-4c59-a167-e59127ef6fab\") " pod="metallb-system/metallb-operator-controller-manager-6977db5dfd-44msc" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.626282 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-65d646cc59-hk6wv"] Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.626893 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-65d646cc59-hk6wv" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.628860 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.628867 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.628952 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-62qsk" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.701435 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-65d646cc59-hk6wv"] Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.710983 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6977db5dfd-44msc" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.789100 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2fcdfd68-9b08-423f-80ec-53032016fdfe-webhook-cert\") pod \"metallb-operator-webhook-server-65d646cc59-hk6wv\" (UID: \"2fcdfd68-9b08-423f-80ec-53032016fdfe\") " pod="metallb-system/metallb-operator-webhook-server-65d646cc59-hk6wv" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.789157 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2fcdfd68-9b08-423f-80ec-53032016fdfe-apiservice-cert\") pod \"metallb-operator-webhook-server-65d646cc59-hk6wv\" (UID: \"2fcdfd68-9b08-423f-80ec-53032016fdfe\") " pod="metallb-system/metallb-operator-webhook-server-65d646cc59-hk6wv" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.789184 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cktt5\" (UniqueName: \"kubernetes.io/projected/2fcdfd68-9b08-423f-80ec-53032016fdfe-kube-api-access-cktt5\") pod \"metallb-operator-webhook-server-65d646cc59-hk6wv\" (UID: \"2fcdfd68-9b08-423f-80ec-53032016fdfe\") " pod="metallb-system/metallb-operator-webhook-server-65d646cc59-hk6wv" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.889871 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2fcdfd68-9b08-423f-80ec-53032016fdfe-webhook-cert\") pod \"metallb-operator-webhook-server-65d646cc59-hk6wv\" (UID: \"2fcdfd68-9b08-423f-80ec-53032016fdfe\") " pod="metallb-system/metallb-operator-webhook-server-65d646cc59-hk6wv" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.889917 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2fcdfd68-9b08-423f-80ec-53032016fdfe-apiservice-cert\") pod \"metallb-operator-webhook-server-65d646cc59-hk6wv\" (UID: \"2fcdfd68-9b08-423f-80ec-53032016fdfe\") " pod="metallb-system/metallb-operator-webhook-server-65d646cc59-hk6wv" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.889941 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cktt5\" (UniqueName: \"kubernetes.io/projected/2fcdfd68-9b08-423f-80ec-53032016fdfe-kube-api-access-cktt5\") pod \"metallb-operator-webhook-server-65d646cc59-hk6wv\" (UID: \"2fcdfd68-9b08-423f-80ec-53032016fdfe\") " pod="metallb-system/metallb-operator-webhook-server-65d646cc59-hk6wv" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.893236 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2fcdfd68-9b08-423f-80ec-53032016fdfe-webhook-cert\") pod \"metallb-operator-webhook-server-65d646cc59-hk6wv\" (UID: \"2fcdfd68-9b08-423f-80ec-53032016fdfe\") " pod="metallb-system/metallb-operator-webhook-server-65d646cc59-hk6wv" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.905335 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2fcdfd68-9b08-423f-80ec-53032016fdfe-apiservice-cert\") pod \"metallb-operator-webhook-server-65d646cc59-hk6wv\" (UID: \"2fcdfd68-9b08-423f-80ec-53032016fdfe\") " pod="metallb-system/metallb-operator-webhook-server-65d646cc59-hk6wv" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.909949 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cktt5\" (UniqueName: \"kubernetes.io/projected/2fcdfd68-9b08-423f-80ec-53032016fdfe-kube-api-access-cktt5\") pod \"metallb-operator-webhook-server-65d646cc59-hk6wv\" (UID: \"2fcdfd68-9b08-423f-80ec-53032016fdfe\") " pod="metallb-system/metallb-operator-webhook-server-65d646cc59-hk6wv" Dec 06 15:57:51 crc kubenswrapper[4813]: I1206 15:57:51.944570 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-65d646cc59-hk6wv" Dec 06 15:57:52 crc kubenswrapper[4813]: I1206 15:57:52.047342 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6977db5dfd-44msc"] Dec 06 15:57:52 crc kubenswrapper[4813]: I1206 15:57:52.128659 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6977db5dfd-44msc" event={"ID":"e7537632-d133-4c59-a167-e59127ef6fab","Type":"ContainerStarted","Data":"f50788da96780d1154de4731bdaf9c86fff4bcd1a86c6b7a24a8245c9721aa31"} Dec 06 15:57:52 crc kubenswrapper[4813]: I1206 15:57:52.237108 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-65d646cc59-hk6wv"] Dec 06 15:57:52 crc kubenswrapper[4813]: W1206 15:57:52.248630 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2fcdfd68_9b08_423f_80ec_53032016fdfe.slice/crio-27690294038ae680ab63d08280e76590af8303a92e18a30692b62fa61138939d WatchSource:0}: Error finding container 27690294038ae680ab63d08280e76590af8303a92e18a30692b62fa61138939d: Status 404 returned error can't find the container with id 27690294038ae680ab63d08280e76590af8303a92e18a30692b62fa61138939d Dec 06 15:57:53 crc kubenswrapper[4813]: I1206 15:57:53.137701 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-65d646cc59-hk6wv" event={"ID":"2fcdfd68-9b08-423f-80ec-53032016fdfe","Type":"ContainerStarted","Data":"27690294038ae680ab63d08280e76590af8303a92e18a30692b62fa61138939d"} Dec 06 15:57:58 crc kubenswrapper[4813]: I1206 15:57:58.171089 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6977db5dfd-44msc" event={"ID":"e7537632-d133-4c59-a167-e59127ef6fab","Type":"ContainerStarted","Data":"ced2a2a796befea06aa84722d571df4f55b25715aea8e8c4774408ef47258056"} Dec 06 15:57:58 crc kubenswrapper[4813]: I1206 15:57:58.171474 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6977db5dfd-44msc" Dec 06 15:57:58 crc kubenswrapper[4813]: I1206 15:57:58.172943 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-65d646cc59-hk6wv" event={"ID":"2fcdfd68-9b08-423f-80ec-53032016fdfe","Type":"ContainerStarted","Data":"1ae45fc7290d6c9eb99e24acec87f7227096cf5ca39010702500ee12c8ccf500"} Dec 06 15:57:58 crc kubenswrapper[4813]: I1206 15:57:58.173189 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-65d646cc59-hk6wv" Dec 06 15:57:58 crc kubenswrapper[4813]: I1206 15:57:58.206153 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-6977db5dfd-44msc" podStartSLOduration=1.6725826910000001 podStartE2EDuration="7.206126326s" podCreationTimestamp="2025-12-06 15:57:51 +0000 UTC" firstStartedPulling="2025-12-06 15:57:52.071768105 +0000 UTC m=+711.962647671" lastFinishedPulling="2025-12-06 15:57:57.60531173 +0000 UTC m=+717.496191306" observedRunningTime="2025-12-06 15:57:58.195879118 +0000 UTC m=+718.086758704" watchObservedRunningTime="2025-12-06 15:57:58.206126326 +0000 UTC m=+718.097005942" Dec 06 15:57:58 crc kubenswrapper[4813]: I1206 15:57:58.228783 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-65d646cc59-hk6wv" podStartSLOduration=1.862065638 podStartE2EDuration="7.228766763s" podCreationTimestamp="2025-12-06 15:57:51 +0000 UTC" firstStartedPulling="2025-12-06 15:57:52.25202456 +0000 UTC m=+712.142904136" lastFinishedPulling="2025-12-06 15:57:57.618725685 +0000 UTC m=+717.509605261" observedRunningTime="2025-12-06 15:57:58.224026509 +0000 UTC m=+718.114906125" watchObservedRunningTime="2025-12-06 15:57:58.228766763 +0000 UTC m=+718.119646349" Dec 06 15:58:11 crc kubenswrapper[4813]: I1206 15:58:11.959567 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-65d646cc59-hk6wv" Dec 06 15:58:31 crc kubenswrapper[4813]: I1206 15:58:31.714758 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6977db5dfd-44msc" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.482932 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-tdbkn"] Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.485867 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-tdbkn" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.487865 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.488603 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.489742 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-9fcv7" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.492798 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-wsvkk"] Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.493459 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wsvkk" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.495456 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.504921 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-wsvkk"] Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.567950 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/74e697b0-ef6a-45c2-8671-c67725e42ab6-reloader\") pod \"frr-k8s-tdbkn\" (UID: \"74e697b0-ef6a-45c2-8671-c67725e42ab6\") " pod="metallb-system/frr-k8s-tdbkn" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.567997 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z698t\" (UniqueName: \"kubernetes.io/projected/a5c26c26-14a5-406c-882b-2950fd2855ed-kube-api-access-z698t\") pod \"frr-k8s-webhook-server-7fcb986d4-wsvkk\" (UID: \"a5c26c26-14a5-406c-882b-2950fd2855ed\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wsvkk" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.568030 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/74e697b0-ef6a-45c2-8671-c67725e42ab6-metrics\") pod \"frr-k8s-tdbkn\" (UID: \"74e697b0-ef6a-45c2-8671-c67725e42ab6\") " pod="metallb-system/frr-k8s-tdbkn" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.568048 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/74e697b0-ef6a-45c2-8671-c67725e42ab6-frr-conf\") pod \"frr-k8s-tdbkn\" (UID: \"74e697b0-ef6a-45c2-8671-c67725e42ab6\") " pod="metallb-system/frr-k8s-tdbkn" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.568076 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zg8t\" (UniqueName: \"kubernetes.io/projected/74e697b0-ef6a-45c2-8671-c67725e42ab6-kube-api-access-9zg8t\") pod \"frr-k8s-tdbkn\" (UID: \"74e697b0-ef6a-45c2-8671-c67725e42ab6\") " pod="metallb-system/frr-k8s-tdbkn" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.568092 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/74e697b0-ef6a-45c2-8671-c67725e42ab6-metrics-certs\") pod \"frr-k8s-tdbkn\" (UID: \"74e697b0-ef6a-45c2-8671-c67725e42ab6\") " pod="metallb-system/frr-k8s-tdbkn" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.568108 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a5c26c26-14a5-406c-882b-2950fd2855ed-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-wsvkk\" (UID: \"a5c26c26-14a5-406c-882b-2950fd2855ed\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wsvkk" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.568124 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/74e697b0-ef6a-45c2-8671-c67725e42ab6-frr-startup\") pod \"frr-k8s-tdbkn\" (UID: \"74e697b0-ef6a-45c2-8671-c67725e42ab6\") " pod="metallb-system/frr-k8s-tdbkn" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.568158 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/74e697b0-ef6a-45c2-8671-c67725e42ab6-frr-sockets\") pod \"frr-k8s-tdbkn\" (UID: \"74e697b0-ef6a-45c2-8671-c67725e42ab6\") " pod="metallb-system/frr-k8s-tdbkn" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.582591 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-ccll7"] Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.583407 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-ccll7" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.586110 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.586293 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.586429 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-gfhtm" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.586643 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.600346 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-mfvl4"] Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.601216 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-mfvl4" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.604698 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.649980 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-mfvl4"] Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.669641 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/45152d05-2dee-4bba-bc56-2e378d0edb9a-memberlist\") pod \"speaker-ccll7\" (UID: \"45152d05-2dee-4bba-bc56-2e378d0edb9a\") " pod="metallb-system/speaker-ccll7" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.669688 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/74e697b0-ef6a-45c2-8671-c67725e42ab6-metrics\") pod \"frr-k8s-tdbkn\" (UID: \"74e697b0-ef6a-45c2-8671-c67725e42ab6\") " pod="metallb-system/frr-k8s-tdbkn" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.669711 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/74e697b0-ef6a-45c2-8671-c67725e42ab6-frr-conf\") pod \"frr-k8s-tdbkn\" (UID: \"74e697b0-ef6a-45c2-8671-c67725e42ab6\") " pod="metallb-system/frr-k8s-tdbkn" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.669840 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a8a9650e-6412-4d98-bdf5-6b57f014eb64-cert\") pod \"controller-f8648f98b-mfvl4\" (UID: \"a8a9650e-6412-4d98-bdf5-6b57f014eb64\") " pod="metallb-system/controller-f8648f98b-mfvl4" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.669935 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/74e697b0-ef6a-45c2-8671-c67725e42ab6-metrics-certs\") pod \"frr-k8s-tdbkn\" (UID: \"74e697b0-ef6a-45c2-8671-c67725e42ab6\") " pod="metallb-system/frr-k8s-tdbkn" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.669956 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zg8t\" (UniqueName: \"kubernetes.io/projected/74e697b0-ef6a-45c2-8671-c67725e42ab6-kube-api-access-9zg8t\") pod \"frr-k8s-tdbkn\" (UID: \"74e697b0-ef6a-45c2-8671-c67725e42ab6\") " pod="metallb-system/frr-k8s-tdbkn" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.669991 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a5c26c26-14a5-406c-882b-2950fd2855ed-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-wsvkk\" (UID: \"a5c26c26-14a5-406c-882b-2950fd2855ed\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wsvkk" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.670015 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a8a9650e-6412-4d98-bdf5-6b57f014eb64-metrics-certs\") pod \"controller-f8648f98b-mfvl4\" (UID: \"a8a9650e-6412-4d98-bdf5-6b57f014eb64\") " pod="metallb-system/controller-f8648f98b-mfvl4" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.670048 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/74e697b0-ef6a-45c2-8671-c67725e42ab6-frr-conf\") pod \"frr-k8s-tdbkn\" (UID: \"74e697b0-ef6a-45c2-8671-c67725e42ab6\") " pod="metallb-system/frr-k8s-tdbkn" Dec 06 15:58:32 crc kubenswrapper[4813]: E1206 15:58:32.670128 4813 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.670152 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/74e697b0-ef6a-45c2-8671-c67725e42ab6-metrics\") pod \"frr-k8s-tdbkn\" (UID: \"74e697b0-ef6a-45c2-8671-c67725e42ab6\") " pod="metallb-system/frr-k8s-tdbkn" Dec 06 15:58:32 crc kubenswrapper[4813]: E1206 15:58:32.670173 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a5c26c26-14a5-406c-882b-2950fd2855ed-cert podName:a5c26c26-14a5-406c-882b-2950fd2855ed nodeName:}" failed. No retries permitted until 2025-12-06 15:58:33.170156642 +0000 UTC m=+753.061036218 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a5c26c26-14a5-406c-882b-2950fd2855ed-cert") pod "frr-k8s-webhook-server-7fcb986d4-wsvkk" (UID: "a5c26c26-14a5-406c-882b-2950fd2855ed") : secret "frr-k8s-webhook-server-cert" not found Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.670438 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpdps\" (UniqueName: \"kubernetes.io/projected/a8a9650e-6412-4d98-bdf5-6b57f014eb64-kube-api-access-jpdps\") pod \"controller-f8648f98b-mfvl4\" (UID: \"a8a9650e-6412-4d98-bdf5-6b57f014eb64\") " pod="metallb-system/controller-f8648f98b-mfvl4" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.670482 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/74e697b0-ef6a-45c2-8671-c67725e42ab6-frr-startup\") pod \"frr-k8s-tdbkn\" (UID: \"74e697b0-ef6a-45c2-8671-c67725e42ab6\") " pod="metallb-system/frr-k8s-tdbkn" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.670512 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/45152d05-2dee-4bba-bc56-2e378d0edb9a-metallb-excludel2\") pod \"speaker-ccll7\" (UID: \"45152d05-2dee-4bba-bc56-2e378d0edb9a\") " pod="metallb-system/speaker-ccll7" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.671232 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/74e697b0-ef6a-45c2-8671-c67725e42ab6-frr-startup\") pod \"frr-k8s-tdbkn\" (UID: \"74e697b0-ef6a-45c2-8671-c67725e42ab6\") " pod="metallb-system/frr-k8s-tdbkn" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.671403 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/74e697b0-ef6a-45c2-8671-c67725e42ab6-frr-sockets\") pod \"frr-k8s-tdbkn\" (UID: \"74e697b0-ef6a-45c2-8671-c67725e42ab6\") " pod="metallb-system/frr-k8s-tdbkn" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.671618 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/74e697b0-ef6a-45c2-8671-c67725e42ab6-frr-sockets\") pod \"frr-k8s-tdbkn\" (UID: \"74e697b0-ef6a-45c2-8671-c67725e42ab6\") " pod="metallb-system/frr-k8s-tdbkn" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.671442 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhd7r\" (UniqueName: \"kubernetes.io/projected/45152d05-2dee-4bba-bc56-2e378d0edb9a-kube-api-access-lhd7r\") pod \"speaker-ccll7\" (UID: \"45152d05-2dee-4bba-bc56-2e378d0edb9a\") " pod="metallb-system/speaker-ccll7" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.671665 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/74e697b0-ef6a-45c2-8671-c67725e42ab6-reloader\") pod \"frr-k8s-tdbkn\" (UID: \"74e697b0-ef6a-45c2-8671-c67725e42ab6\") " pod="metallb-system/frr-k8s-tdbkn" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.671718 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z698t\" (UniqueName: \"kubernetes.io/projected/a5c26c26-14a5-406c-882b-2950fd2855ed-kube-api-access-z698t\") pod \"frr-k8s-webhook-server-7fcb986d4-wsvkk\" (UID: \"a5c26c26-14a5-406c-882b-2950fd2855ed\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wsvkk" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.671734 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/45152d05-2dee-4bba-bc56-2e378d0edb9a-metrics-certs\") pod \"speaker-ccll7\" (UID: \"45152d05-2dee-4bba-bc56-2e378d0edb9a\") " pod="metallb-system/speaker-ccll7" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.671927 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/74e697b0-ef6a-45c2-8671-c67725e42ab6-reloader\") pod \"frr-k8s-tdbkn\" (UID: \"74e697b0-ef6a-45c2-8671-c67725e42ab6\") " pod="metallb-system/frr-k8s-tdbkn" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.678711 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/74e697b0-ef6a-45c2-8671-c67725e42ab6-metrics-certs\") pod \"frr-k8s-tdbkn\" (UID: \"74e697b0-ef6a-45c2-8671-c67725e42ab6\") " pod="metallb-system/frr-k8s-tdbkn" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.684144 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zg8t\" (UniqueName: \"kubernetes.io/projected/74e697b0-ef6a-45c2-8671-c67725e42ab6-kube-api-access-9zg8t\") pod \"frr-k8s-tdbkn\" (UID: \"74e697b0-ef6a-45c2-8671-c67725e42ab6\") " pod="metallb-system/frr-k8s-tdbkn" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.687601 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z698t\" (UniqueName: \"kubernetes.io/projected/a5c26c26-14a5-406c-882b-2950fd2855ed-kube-api-access-z698t\") pod \"frr-k8s-webhook-server-7fcb986d4-wsvkk\" (UID: \"a5c26c26-14a5-406c-882b-2950fd2855ed\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wsvkk" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.773321 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhd7r\" (UniqueName: \"kubernetes.io/projected/45152d05-2dee-4bba-bc56-2e378d0edb9a-kube-api-access-lhd7r\") pod \"speaker-ccll7\" (UID: \"45152d05-2dee-4bba-bc56-2e378d0edb9a\") " pod="metallb-system/speaker-ccll7" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.773371 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/45152d05-2dee-4bba-bc56-2e378d0edb9a-metrics-certs\") pod \"speaker-ccll7\" (UID: \"45152d05-2dee-4bba-bc56-2e378d0edb9a\") " pod="metallb-system/speaker-ccll7" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.773420 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/45152d05-2dee-4bba-bc56-2e378d0edb9a-memberlist\") pod \"speaker-ccll7\" (UID: \"45152d05-2dee-4bba-bc56-2e378d0edb9a\") " pod="metallb-system/speaker-ccll7" Dec 06 15:58:32 crc kubenswrapper[4813]: E1206 15:58:32.773671 4813 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 06 15:58:32 crc kubenswrapper[4813]: E1206 15:58:32.773751 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/45152d05-2dee-4bba-bc56-2e378d0edb9a-memberlist podName:45152d05-2dee-4bba-bc56-2e378d0edb9a nodeName:}" failed. No retries permitted until 2025-12-06 15:58:33.273732626 +0000 UTC m=+753.164612202 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/45152d05-2dee-4bba-bc56-2e378d0edb9a-memberlist") pod "speaker-ccll7" (UID: "45152d05-2dee-4bba-bc56-2e378d0edb9a") : secret "metallb-memberlist" not found Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.773893 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a8a9650e-6412-4d98-bdf5-6b57f014eb64-cert\") pod \"controller-f8648f98b-mfvl4\" (UID: \"a8a9650e-6412-4d98-bdf5-6b57f014eb64\") " pod="metallb-system/controller-f8648f98b-mfvl4" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.773929 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a8a9650e-6412-4d98-bdf5-6b57f014eb64-metrics-certs\") pod \"controller-f8648f98b-mfvl4\" (UID: \"a8a9650e-6412-4d98-bdf5-6b57f014eb64\") " pod="metallb-system/controller-f8648f98b-mfvl4" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.773947 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpdps\" (UniqueName: \"kubernetes.io/projected/a8a9650e-6412-4d98-bdf5-6b57f014eb64-kube-api-access-jpdps\") pod \"controller-f8648f98b-mfvl4\" (UID: \"a8a9650e-6412-4d98-bdf5-6b57f014eb64\") " pod="metallb-system/controller-f8648f98b-mfvl4" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.773979 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/45152d05-2dee-4bba-bc56-2e378d0edb9a-metallb-excludel2\") pod \"speaker-ccll7\" (UID: \"45152d05-2dee-4bba-bc56-2e378d0edb9a\") " pod="metallb-system/speaker-ccll7" Dec 06 15:58:32 crc kubenswrapper[4813]: E1206 15:58:32.774092 4813 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Dec 06 15:58:32 crc kubenswrapper[4813]: E1206 15:58:32.774170 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a8a9650e-6412-4d98-bdf5-6b57f014eb64-metrics-certs podName:a8a9650e-6412-4d98-bdf5-6b57f014eb64 nodeName:}" failed. No retries permitted until 2025-12-06 15:58:33.274151206 +0000 UTC m=+753.165030782 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a8a9650e-6412-4d98-bdf5-6b57f014eb64-metrics-certs") pod "controller-f8648f98b-mfvl4" (UID: "a8a9650e-6412-4d98-bdf5-6b57f014eb64") : secret "controller-certs-secret" not found Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.774567 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/45152d05-2dee-4bba-bc56-2e378d0edb9a-metallb-excludel2\") pod \"speaker-ccll7\" (UID: \"45152d05-2dee-4bba-bc56-2e378d0edb9a\") " pod="metallb-system/speaker-ccll7" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.777394 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.777574 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/45152d05-2dee-4bba-bc56-2e378d0edb9a-metrics-certs\") pod \"speaker-ccll7\" (UID: \"45152d05-2dee-4bba-bc56-2e378d0edb9a\") " pod="metallb-system/speaker-ccll7" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.786567 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a8a9650e-6412-4d98-bdf5-6b57f014eb64-cert\") pod \"controller-f8648f98b-mfvl4\" (UID: \"a8a9650e-6412-4d98-bdf5-6b57f014eb64\") " pod="metallb-system/controller-f8648f98b-mfvl4" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.795886 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhd7r\" (UniqueName: \"kubernetes.io/projected/45152d05-2dee-4bba-bc56-2e378d0edb9a-kube-api-access-lhd7r\") pod \"speaker-ccll7\" (UID: \"45152d05-2dee-4bba-bc56-2e378d0edb9a\") " pod="metallb-system/speaker-ccll7" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.800104 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpdps\" (UniqueName: \"kubernetes.io/projected/a8a9650e-6412-4d98-bdf5-6b57f014eb64-kube-api-access-jpdps\") pod \"controller-f8648f98b-mfvl4\" (UID: \"a8a9650e-6412-4d98-bdf5-6b57f014eb64\") " pod="metallb-system/controller-f8648f98b-mfvl4" Dec 06 15:58:32 crc kubenswrapper[4813]: I1206 15:58:32.805072 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-tdbkn" Dec 06 15:58:33 crc kubenswrapper[4813]: I1206 15:58:33.180017 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a5c26c26-14a5-406c-882b-2950fd2855ed-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-wsvkk\" (UID: \"a5c26c26-14a5-406c-882b-2950fd2855ed\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wsvkk" Dec 06 15:58:33 crc kubenswrapper[4813]: I1206 15:58:33.185139 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a5c26c26-14a5-406c-882b-2950fd2855ed-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-wsvkk\" (UID: \"a5c26c26-14a5-406c-882b-2950fd2855ed\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wsvkk" Dec 06 15:58:33 crc kubenswrapper[4813]: I1206 15:58:33.282809 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a8a9650e-6412-4d98-bdf5-6b57f014eb64-metrics-certs\") pod \"controller-f8648f98b-mfvl4\" (UID: \"a8a9650e-6412-4d98-bdf5-6b57f014eb64\") " pod="metallb-system/controller-f8648f98b-mfvl4" Dec 06 15:58:33 crc kubenswrapper[4813]: I1206 15:58:33.282987 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/45152d05-2dee-4bba-bc56-2e378d0edb9a-memberlist\") pod \"speaker-ccll7\" (UID: \"45152d05-2dee-4bba-bc56-2e378d0edb9a\") " pod="metallb-system/speaker-ccll7" Dec 06 15:58:33 crc kubenswrapper[4813]: E1206 15:58:33.283209 4813 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 06 15:58:33 crc kubenswrapper[4813]: E1206 15:58:33.283357 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/45152d05-2dee-4bba-bc56-2e378d0edb9a-memberlist podName:45152d05-2dee-4bba-bc56-2e378d0edb9a nodeName:}" failed. No retries permitted until 2025-12-06 15:58:34.283334056 +0000 UTC m=+754.174213632 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/45152d05-2dee-4bba-bc56-2e378d0edb9a-memberlist") pod "speaker-ccll7" (UID: "45152d05-2dee-4bba-bc56-2e378d0edb9a") : secret "metallb-memberlist" not found Dec 06 15:58:33 crc kubenswrapper[4813]: I1206 15:58:33.287934 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a8a9650e-6412-4d98-bdf5-6b57f014eb64-metrics-certs\") pod \"controller-f8648f98b-mfvl4\" (UID: \"a8a9650e-6412-4d98-bdf5-6b57f014eb64\") " pod="metallb-system/controller-f8648f98b-mfvl4" Dec 06 15:58:33 crc kubenswrapper[4813]: I1206 15:58:33.389943 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tdbkn" event={"ID":"74e697b0-ef6a-45c2-8671-c67725e42ab6","Type":"ContainerStarted","Data":"da0d7dc7d9b602a816f72ee063c5b8bb240013b413d5664ac80643af552f9fd5"} Dec 06 15:58:33 crc kubenswrapper[4813]: I1206 15:58:33.415680 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wsvkk" Dec 06 15:58:33 crc kubenswrapper[4813]: I1206 15:58:33.512899 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-mfvl4" Dec 06 15:58:33 crc kubenswrapper[4813]: I1206 15:58:33.739219 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-wsvkk"] Dec 06 15:58:33 crc kubenswrapper[4813]: I1206 15:58:33.784203 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-mfvl4"] Dec 06 15:58:34 crc kubenswrapper[4813]: I1206 15:58:34.307158 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/45152d05-2dee-4bba-bc56-2e378d0edb9a-memberlist\") pod \"speaker-ccll7\" (UID: \"45152d05-2dee-4bba-bc56-2e378d0edb9a\") " pod="metallb-system/speaker-ccll7" Dec 06 15:58:34 crc kubenswrapper[4813]: I1206 15:58:34.316931 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/45152d05-2dee-4bba-bc56-2e378d0edb9a-memberlist\") pod \"speaker-ccll7\" (UID: \"45152d05-2dee-4bba-bc56-2e378d0edb9a\") " pod="metallb-system/speaker-ccll7" Dec 06 15:58:34 crc kubenswrapper[4813]: I1206 15:58:34.401309 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-ccll7" Dec 06 15:58:34 crc kubenswrapper[4813]: I1206 15:58:34.412416 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wsvkk" event={"ID":"a5c26c26-14a5-406c-882b-2950fd2855ed","Type":"ContainerStarted","Data":"5ee0379a494532ad7df1d52118a317052ce055b7c890ac2913125350b92b0945"} Dec 06 15:58:34 crc kubenswrapper[4813]: I1206 15:58:34.418831 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-mfvl4" event={"ID":"a8a9650e-6412-4d98-bdf5-6b57f014eb64","Type":"ContainerStarted","Data":"40f9e61f59f9c81b63ae1722eba2a536f12f851c80e22d4f60fab76903c27da2"} Dec 06 15:58:34 crc kubenswrapper[4813]: I1206 15:58:34.418872 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-mfvl4" event={"ID":"a8a9650e-6412-4d98-bdf5-6b57f014eb64","Type":"ContainerStarted","Data":"57510cd38a14b3ed62a9a99fd40910782a9dd6ad8b1341d00c7a787e22369166"} Dec 06 15:58:34 crc kubenswrapper[4813]: I1206 15:58:34.418882 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-mfvl4" event={"ID":"a8a9650e-6412-4d98-bdf5-6b57f014eb64","Type":"ContainerStarted","Data":"802797304b37dc1b06eb8f4637d2aa2caae3d61f9174b0ffb0309c13808d070c"} Dec 06 15:58:34 crc kubenswrapper[4813]: I1206 15:58:34.419357 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-mfvl4" Dec 06 15:58:34 crc kubenswrapper[4813]: I1206 15:58:34.439672 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-mfvl4" podStartSLOduration=2.439649271 podStartE2EDuration="2.439649271s" podCreationTimestamp="2025-12-06 15:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:58:34.439023196 +0000 UTC m=+754.329902772" watchObservedRunningTime="2025-12-06 15:58:34.439649271 +0000 UTC m=+754.330528847" Dec 06 15:58:35 crc kubenswrapper[4813]: I1206 15:58:35.425874 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-ccll7" event={"ID":"45152d05-2dee-4bba-bc56-2e378d0edb9a","Type":"ContainerStarted","Data":"76295a2508723b7dfbbe35015153084099ea366ec50b1275995424ab78522af7"} Dec 06 15:58:35 crc kubenswrapper[4813]: I1206 15:58:35.426086 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-ccll7" event={"ID":"45152d05-2dee-4bba-bc56-2e378d0edb9a","Type":"ContainerStarted","Data":"e530217a29e759fc909b8f648602bd56a671a5220b266b030e662e0e1ffa2ab1"} Dec 06 15:58:35 crc kubenswrapper[4813]: I1206 15:58:35.426095 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-ccll7" event={"ID":"45152d05-2dee-4bba-bc56-2e378d0edb9a","Type":"ContainerStarted","Data":"2b41d1dd3dbf5cb46b6fa130244eb85a3cbd5771aca9ac046aac5179896db876"} Dec 06 15:58:35 crc kubenswrapper[4813]: I1206 15:58:35.426617 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-ccll7" Dec 06 15:58:35 crc kubenswrapper[4813]: I1206 15:58:35.444795 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-ccll7" podStartSLOduration=3.44472232 podStartE2EDuration="3.44472232s" podCreationTimestamp="2025-12-06 15:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 15:58:35.441427371 +0000 UTC m=+755.332306947" watchObservedRunningTime="2025-12-06 15:58:35.44472232 +0000 UTC m=+755.335601896" Dec 06 15:58:38 crc kubenswrapper[4813]: I1206 15:58:38.870915 4813 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 06 15:58:40 crc kubenswrapper[4813]: I1206 15:58:40.459720 4813 generic.go:334] "Generic (PLEG): container finished" podID="74e697b0-ef6a-45c2-8671-c67725e42ab6" containerID="471238e43d04988540bfebf80aae9e91c46e2ebee2af1fe471b20a056e774419" exitCode=0 Dec 06 15:58:40 crc kubenswrapper[4813]: I1206 15:58:40.460179 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tdbkn" event={"ID":"74e697b0-ef6a-45c2-8671-c67725e42ab6","Type":"ContainerDied","Data":"471238e43d04988540bfebf80aae9e91c46e2ebee2af1fe471b20a056e774419"} Dec 06 15:58:40 crc kubenswrapper[4813]: I1206 15:58:40.463792 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wsvkk" event={"ID":"a5c26c26-14a5-406c-882b-2950fd2855ed","Type":"ContainerStarted","Data":"933471fc6016f4cffebdc53837ba53a32444f0e310c374bc71172df6dd62a820"} Dec 06 15:58:40 crc kubenswrapper[4813]: I1206 15:58:40.464158 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wsvkk" Dec 06 15:58:40 crc kubenswrapper[4813]: I1206 15:58:40.512483 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wsvkk" podStartSLOduration=2.009102333 podStartE2EDuration="8.512461448s" podCreationTimestamp="2025-12-06 15:58:32 +0000 UTC" firstStartedPulling="2025-12-06 15:58:33.75958805 +0000 UTC m=+753.650467626" lastFinishedPulling="2025-12-06 15:58:40.262947165 +0000 UTC m=+760.153826741" observedRunningTime="2025-12-06 15:58:40.507653521 +0000 UTC m=+760.398533097" watchObservedRunningTime="2025-12-06 15:58:40.512461448 +0000 UTC m=+760.403341034" Dec 06 15:58:41 crc kubenswrapper[4813]: I1206 15:58:41.471015 4813 generic.go:334] "Generic (PLEG): container finished" podID="74e697b0-ef6a-45c2-8671-c67725e42ab6" containerID="11ffee52e013951aabd2e3c19c3f7345618c30bb9354b4f93894df3f9e4cc325" exitCode=0 Dec 06 15:58:41 crc kubenswrapper[4813]: I1206 15:58:41.471602 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tdbkn" event={"ID":"74e697b0-ef6a-45c2-8671-c67725e42ab6","Type":"ContainerDied","Data":"11ffee52e013951aabd2e3c19c3f7345618c30bb9354b4f93894df3f9e4cc325"} Dec 06 15:58:42 crc kubenswrapper[4813]: I1206 15:58:42.482067 4813 generic.go:334] "Generic (PLEG): container finished" podID="74e697b0-ef6a-45c2-8671-c67725e42ab6" containerID="44fa2245467c1b5ddc8eb548a073bf62dfde149988d4e27d930fc0aa1a625ba1" exitCode=0 Dec 06 15:58:42 crc kubenswrapper[4813]: I1206 15:58:42.482421 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tdbkn" event={"ID":"74e697b0-ef6a-45c2-8671-c67725e42ab6","Type":"ContainerDied","Data":"44fa2245467c1b5ddc8eb548a073bf62dfde149988d4e27d930fc0aa1a625ba1"} Dec 06 15:58:43 crc kubenswrapper[4813]: I1206 15:58:43.494290 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tdbkn" event={"ID":"74e697b0-ef6a-45c2-8671-c67725e42ab6","Type":"ContainerStarted","Data":"02f0a6e9179bf9775fbb50b6881320a213e83c5ea423d0772b6328ff074d44b7"} Dec 06 15:58:43 crc kubenswrapper[4813]: I1206 15:58:43.494594 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tdbkn" event={"ID":"74e697b0-ef6a-45c2-8671-c67725e42ab6","Type":"ContainerStarted","Data":"8cc202dfe7420026970d28576820755e5a1a066b9755cd35b343091fb0b5e40a"} Dec 06 15:58:43 crc kubenswrapper[4813]: I1206 15:58:43.494616 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tdbkn" event={"ID":"74e697b0-ef6a-45c2-8671-c67725e42ab6","Type":"ContainerStarted","Data":"2e40533b931727b90fe3abe1084bf108f19b716d28d113ee879d514c00659c31"} Dec 06 15:58:43 crc kubenswrapper[4813]: I1206 15:58:43.494635 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tdbkn" event={"ID":"74e697b0-ef6a-45c2-8671-c67725e42ab6","Type":"ContainerStarted","Data":"00502b0d73de2a42edba50106fb45fa9e733500125ea2e93ae19e1a68de67dac"} Dec 06 15:58:43 crc kubenswrapper[4813]: I1206 15:58:43.521590 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-mfvl4" Dec 06 15:58:44 crc kubenswrapper[4813]: I1206 15:58:44.405499 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-ccll7" Dec 06 15:58:44 crc kubenswrapper[4813]: I1206 15:58:44.508225 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tdbkn" event={"ID":"74e697b0-ef6a-45c2-8671-c67725e42ab6","Type":"ContainerStarted","Data":"993b68692a949908efb19dd6fe9f9d4770299f7e6437cf607b5a64340cb47f13"} Dec 06 15:58:44 crc kubenswrapper[4813]: I1206 15:58:44.508309 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tdbkn" event={"ID":"74e697b0-ef6a-45c2-8671-c67725e42ab6","Type":"ContainerStarted","Data":"e8f7b1278f32e08b17fbeceaa219fa2d1fc2aa9aea40cf42515f912a4618a328"} Dec 06 15:58:44 crc kubenswrapper[4813]: I1206 15:58:44.508431 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-tdbkn" Dec 06 15:58:44 crc kubenswrapper[4813]: I1206 15:58:44.541513 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-tdbkn" podStartSLOduration=5.205540769 podStartE2EDuration="12.541495414s" podCreationTimestamp="2025-12-06 15:58:32 +0000 UTC" firstStartedPulling="2025-12-06 15:58:32.907853528 +0000 UTC m=+752.798733104" lastFinishedPulling="2025-12-06 15:58:40.243808163 +0000 UTC m=+760.134687749" observedRunningTime="2025-12-06 15:58:44.539183878 +0000 UTC m=+764.430063484" watchObservedRunningTime="2025-12-06 15:58:44.541495414 +0000 UTC m=+764.432375000" Dec 06 15:58:47 crc kubenswrapper[4813]: I1206 15:58:47.345773 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-dcpm7"] Dec 06 15:58:47 crc kubenswrapper[4813]: I1206 15:58:47.346757 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-dcpm7" Dec 06 15:58:47 crc kubenswrapper[4813]: I1206 15:58:47.349730 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-5cc9l" Dec 06 15:58:47 crc kubenswrapper[4813]: I1206 15:58:47.349968 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 06 15:58:47 crc kubenswrapper[4813]: I1206 15:58:47.350172 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 06 15:58:47 crc kubenswrapper[4813]: I1206 15:58:47.365869 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-dcpm7"] Dec 06 15:58:47 crc kubenswrapper[4813]: I1206 15:58:47.503835 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4hm7\" (UniqueName: \"kubernetes.io/projected/3cefa5c1-a8ed-495e-9781-ef2ec827113e-kube-api-access-z4hm7\") pod \"openstack-operator-index-dcpm7\" (UID: \"3cefa5c1-a8ed-495e-9781-ef2ec827113e\") " pod="openstack-operators/openstack-operator-index-dcpm7" Dec 06 15:58:47 crc kubenswrapper[4813]: I1206 15:58:47.605578 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4hm7\" (UniqueName: \"kubernetes.io/projected/3cefa5c1-a8ed-495e-9781-ef2ec827113e-kube-api-access-z4hm7\") pod \"openstack-operator-index-dcpm7\" (UID: \"3cefa5c1-a8ed-495e-9781-ef2ec827113e\") " pod="openstack-operators/openstack-operator-index-dcpm7" Dec 06 15:58:47 crc kubenswrapper[4813]: I1206 15:58:47.628501 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4hm7\" (UniqueName: \"kubernetes.io/projected/3cefa5c1-a8ed-495e-9781-ef2ec827113e-kube-api-access-z4hm7\") pod \"openstack-operator-index-dcpm7\" (UID: \"3cefa5c1-a8ed-495e-9781-ef2ec827113e\") " pod="openstack-operators/openstack-operator-index-dcpm7" Dec 06 15:58:47 crc kubenswrapper[4813]: I1206 15:58:47.713650 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-dcpm7" Dec 06 15:58:47 crc kubenswrapper[4813]: I1206 15:58:47.806313 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-tdbkn" Dec 06 15:58:47 crc kubenswrapper[4813]: I1206 15:58:47.870830 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-tdbkn" Dec 06 15:58:47 crc kubenswrapper[4813]: I1206 15:58:47.978229 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-dcpm7"] Dec 06 15:58:48 crc kubenswrapper[4813]: I1206 15:58:48.534297 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-dcpm7" event={"ID":"3cefa5c1-a8ed-495e-9781-ef2ec827113e","Type":"ContainerStarted","Data":"4e16bdf4d798f9b2f40c5eebda3b537963e6193281746c5230a19c56ad3c3a3c"} Dec 06 15:58:49 crc kubenswrapper[4813]: I1206 15:58:49.428317 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 15:58:49 crc kubenswrapper[4813]: I1206 15:58:49.428377 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 15:58:50 crc kubenswrapper[4813]: I1206 15:58:50.696373 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-dcpm7"] Dec 06 15:58:51 crc kubenswrapper[4813]: I1206 15:58:51.309812 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-vgjq7"] Dec 06 15:58:51 crc kubenswrapper[4813]: I1206 15:58:51.311793 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vgjq7" Dec 06 15:58:51 crc kubenswrapper[4813]: I1206 15:58:51.332897 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-vgjq7"] Dec 06 15:58:51 crc kubenswrapper[4813]: I1206 15:58:51.469470 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdxht\" (UniqueName: \"kubernetes.io/projected/74a02e6c-0417-4c92-9edd-18925cfbf96e-kube-api-access-xdxht\") pod \"openstack-operator-index-vgjq7\" (UID: \"74a02e6c-0417-4c92-9edd-18925cfbf96e\") " pod="openstack-operators/openstack-operator-index-vgjq7" Dec 06 15:58:51 crc kubenswrapper[4813]: I1206 15:58:51.557418 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-dcpm7" event={"ID":"3cefa5c1-a8ed-495e-9781-ef2ec827113e","Type":"ContainerStarted","Data":"e52c26813371a59fa7fac5d051835011d26c8160c215fe08eedbc2cc72aa845f"} Dec 06 15:58:51 crc kubenswrapper[4813]: I1206 15:58:51.557596 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-dcpm7" podUID="3cefa5c1-a8ed-495e-9781-ef2ec827113e" containerName="registry-server" containerID="cri-o://e52c26813371a59fa7fac5d051835011d26c8160c215fe08eedbc2cc72aa845f" gracePeriod=2 Dec 06 15:58:51 crc kubenswrapper[4813]: I1206 15:58:51.571075 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdxht\" (UniqueName: \"kubernetes.io/projected/74a02e6c-0417-4c92-9edd-18925cfbf96e-kube-api-access-xdxht\") pod \"openstack-operator-index-vgjq7\" (UID: \"74a02e6c-0417-4c92-9edd-18925cfbf96e\") " pod="openstack-operators/openstack-operator-index-vgjq7" Dec 06 15:58:51 crc kubenswrapper[4813]: I1206 15:58:51.590157 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-dcpm7" podStartSLOduration=1.5987412920000001 podStartE2EDuration="4.590135249s" podCreationTimestamp="2025-12-06 15:58:47 +0000 UTC" firstStartedPulling="2025-12-06 15:58:47.985545667 +0000 UTC m=+767.876425243" lastFinishedPulling="2025-12-06 15:58:50.976939624 +0000 UTC m=+770.867819200" observedRunningTime="2025-12-06 15:58:51.582028222 +0000 UTC m=+771.472907828" watchObservedRunningTime="2025-12-06 15:58:51.590135249 +0000 UTC m=+771.481014855" Dec 06 15:58:51 crc kubenswrapper[4813]: I1206 15:58:51.616073 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdxht\" (UniqueName: \"kubernetes.io/projected/74a02e6c-0417-4c92-9edd-18925cfbf96e-kube-api-access-xdxht\") pod \"openstack-operator-index-vgjq7\" (UID: \"74a02e6c-0417-4c92-9edd-18925cfbf96e\") " pod="openstack-operators/openstack-operator-index-vgjq7" Dec 06 15:58:51 crc kubenswrapper[4813]: I1206 15:58:51.635528 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vgjq7" Dec 06 15:58:51 crc kubenswrapper[4813]: I1206 15:58:51.961099 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-dcpm7" Dec 06 15:58:52 crc kubenswrapper[4813]: I1206 15:58:52.078980 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4hm7\" (UniqueName: \"kubernetes.io/projected/3cefa5c1-a8ed-495e-9781-ef2ec827113e-kube-api-access-z4hm7\") pod \"3cefa5c1-a8ed-495e-9781-ef2ec827113e\" (UID: \"3cefa5c1-a8ed-495e-9781-ef2ec827113e\") " Dec 06 15:58:52 crc kubenswrapper[4813]: I1206 15:58:52.085876 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cefa5c1-a8ed-495e-9781-ef2ec827113e-kube-api-access-z4hm7" (OuterVolumeSpecName: "kube-api-access-z4hm7") pod "3cefa5c1-a8ed-495e-9781-ef2ec827113e" (UID: "3cefa5c1-a8ed-495e-9781-ef2ec827113e"). InnerVolumeSpecName "kube-api-access-z4hm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:58:52 crc kubenswrapper[4813]: I1206 15:58:52.139126 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-vgjq7"] Dec 06 15:58:52 crc kubenswrapper[4813]: I1206 15:58:52.180914 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4hm7\" (UniqueName: \"kubernetes.io/projected/3cefa5c1-a8ed-495e-9781-ef2ec827113e-kube-api-access-z4hm7\") on node \"crc\" DevicePath \"\"" Dec 06 15:58:52 crc kubenswrapper[4813]: I1206 15:58:52.565688 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vgjq7" event={"ID":"74a02e6c-0417-4c92-9edd-18925cfbf96e","Type":"ContainerStarted","Data":"01bced9c0b92cb86fca856c64f523626f72f63aeb3e8eae8d8fac91f1f7cd516"} Dec 06 15:58:52 crc kubenswrapper[4813]: I1206 15:58:52.565977 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vgjq7" event={"ID":"74a02e6c-0417-4c92-9edd-18925cfbf96e","Type":"ContainerStarted","Data":"bc8f1932daf7fa5ea20d0ad70803dd3ec0ebeb7a9a3abb97ecc55a59ff7733e0"} Dec 06 15:58:52 crc kubenswrapper[4813]: I1206 15:58:52.568750 4813 generic.go:334] "Generic (PLEG): container finished" podID="3cefa5c1-a8ed-495e-9781-ef2ec827113e" containerID="e52c26813371a59fa7fac5d051835011d26c8160c215fe08eedbc2cc72aa845f" exitCode=0 Dec 06 15:58:52 crc kubenswrapper[4813]: I1206 15:58:52.568810 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-dcpm7" event={"ID":"3cefa5c1-a8ed-495e-9781-ef2ec827113e","Type":"ContainerDied","Data":"e52c26813371a59fa7fac5d051835011d26c8160c215fe08eedbc2cc72aa845f"} Dec 06 15:58:52 crc kubenswrapper[4813]: I1206 15:58:52.568846 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-dcpm7" event={"ID":"3cefa5c1-a8ed-495e-9781-ef2ec827113e","Type":"ContainerDied","Data":"4e16bdf4d798f9b2f40c5eebda3b537963e6193281746c5230a19c56ad3c3a3c"} Dec 06 15:58:52 crc kubenswrapper[4813]: I1206 15:58:52.568874 4813 scope.go:117] "RemoveContainer" containerID="e52c26813371a59fa7fac5d051835011d26c8160c215fe08eedbc2cc72aa845f" Dec 06 15:58:52 crc kubenswrapper[4813]: I1206 15:58:52.569364 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-dcpm7" Dec 06 15:58:52 crc kubenswrapper[4813]: I1206 15:58:52.601573 4813 scope.go:117] "RemoveContainer" containerID="e52c26813371a59fa7fac5d051835011d26c8160c215fe08eedbc2cc72aa845f" Dec 06 15:58:52 crc kubenswrapper[4813]: E1206 15:58:52.602828 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e52c26813371a59fa7fac5d051835011d26c8160c215fe08eedbc2cc72aa845f\": container with ID starting with e52c26813371a59fa7fac5d051835011d26c8160c215fe08eedbc2cc72aa845f not found: ID does not exist" containerID="e52c26813371a59fa7fac5d051835011d26c8160c215fe08eedbc2cc72aa845f" Dec 06 15:58:52 crc kubenswrapper[4813]: I1206 15:58:52.602874 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e52c26813371a59fa7fac5d051835011d26c8160c215fe08eedbc2cc72aa845f"} err="failed to get container status \"e52c26813371a59fa7fac5d051835011d26c8160c215fe08eedbc2cc72aa845f\": rpc error: code = NotFound desc = could not find container \"e52c26813371a59fa7fac5d051835011d26c8160c215fe08eedbc2cc72aa845f\": container with ID starting with e52c26813371a59fa7fac5d051835011d26c8160c215fe08eedbc2cc72aa845f not found: ID does not exist" Dec 06 15:58:52 crc kubenswrapper[4813]: I1206 15:58:52.614489 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-vgjq7" podStartSLOduration=1.5666806659999999 podStartE2EDuration="1.614466846s" podCreationTimestamp="2025-12-06 15:58:51 +0000 UTC" firstStartedPulling="2025-12-06 15:58:52.147673723 +0000 UTC m=+772.038553299" lastFinishedPulling="2025-12-06 15:58:52.195459893 +0000 UTC m=+772.086339479" observedRunningTime="2025-12-06 15:58:52.596221908 +0000 UTC m=+772.487101494" watchObservedRunningTime="2025-12-06 15:58:52.614466846 +0000 UTC m=+772.505346432" Dec 06 15:58:52 crc kubenswrapper[4813]: I1206 15:58:52.618016 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-dcpm7"] Dec 06 15:58:52 crc kubenswrapper[4813]: I1206 15:58:52.622758 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-dcpm7"] Dec 06 15:58:52 crc kubenswrapper[4813]: I1206 15:58:52.817726 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-tdbkn" Dec 06 15:58:53 crc kubenswrapper[4813]: I1206 15:58:53.421909 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wsvkk" Dec 06 15:58:54 crc kubenswrapper[4813]: I1206 15:58:54.505679 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cefa5c1-a8ed-495e-9781-ef2ec827113e" path="/var/lib/kubelet/pods/3cefa5c1-a8ed-495e-9781-ef2ec827113e/volumes" Dec 06 15:59:01 crc kubenswrapper[4813]: I1206 15:59:01.637050 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-vgjq7" Dec 06 15:59:01 crc kubenswrapper[4813]: I1206 15:59:01.638348 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-vgjq7" Dec 06 15:59:01 crc kubenswrapper[4813]: I1206 15:59:01.678313 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-vgjq7" Dec 06 15:59:01 crc kubenswrapper[4813]: I1206 15:59:01.723851 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-vgjq7" Dec 06 15:59:02 crc kubenswrapper[4813]: I1206 15:59:02.952927 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f"] Dec 06 15:59:02 crc kubenswrapper[4813]: E1206 15:59:02.953561 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cefa5c1-a8ed-495e-9781-ef2ec827113e" containerName="registry-server" Dec 06 15:59:02 crc kubenswrapper[4813]: I1206 15:59:02.953582 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cefa5c1-a8ed-495e-9781-ef2ec827113e" containerName="registry-server" Dec 06 15:59:02 crc kubenswrapper[4813]: I1206 15:59:02.953782 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cefa5c1-a8ed-495e-9781-ef2ec827113e" containerName="registry-server" Dec 06 15:59:02 crc kubenswrapper[4813]: I1206 15:59:02.955123 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f" Dec 06 15:59:02 crc kubenswrapper[4813]: I1206 15:59:02.958631 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-sjvf4" Dec 06 15:59:02 crc kubenswrapper[4813]: I1206 15:59:02.974373 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f"] Dec 06 15:59:03 crc kubenswrapper[4813]: I1206 15:59:03.057943 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtsfr\" (UniqueName: \"kubernetes.io/projected/89c13d20-79f6-4a15-9c75-cf199679fc7e-kube-api-access-xtsfr\") pod \"05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f\" (UID: \"89c13d20-79f6-4a15-9c75-cf199679fc7e\") " pod="openstack-operators/05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f" Dec 06 15:59:03 crc kubenswrapper[4813]: I1206 15:59:03.058033 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/89c13d20-79f6-4a15-9c75-cf199679fc7e-util\") pod \"05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f\" (UID: \"89c13d20-79f6-4a15-9c75-cf199679fc7e\") " pod="openstack-operators/05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f" Dec 06 15:59:03 crc kubenswrapper[4813]: I1206 15:59:03.058108 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/89c13d20-79f6-4a15-9c75-cf199679fc7e-bundle\") pod \"05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f\" (UID: \"89c13d20-79f6-4a15-9c75-cf199679fc7e\") " pod="openstack-operators/05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f" Dec 06 15:59:03 crc kubenswrapper[4813]: I1206 15:59:03.159883 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtsfr\" (UniqueName: \"kubernetes.io/projected/89c13d20-79f6-4a15-9c75-cf199679fc7e-kube-api-access-xtsfr\") pod \"05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f\" (UID: \"89c13d20-79f6-4a15-9c75-cf199679fc7e\") " pod="openstack-operators/05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f" Dec 06 15:59:03 crc kubenswrapper[4813]: I1206 15:59:03.159972 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/89c13d20-79f6-4a15-9c75-cf199679fc7e-util\") pod \"05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f\" (UID: \"89c13d20-79f6-4a15-9c75-cf199679fc7e\") " pod="openstack-operators/05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f" Dec 06 15:59:03 crc kubenswrapper[4813]: I1206 15:59:03.160043 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/89c13d20-79f6-4a15-9c75-cf199679fc7e-bundle\") pod \"05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f\" (UID: \"89c13d20-79f6-4a15-9c75-cf199679fc7e\") " pod="openstack-operators/05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f" Dec 06 15:59:03 crc kubenswrapper[4813]: I1206 15:59:03.160801 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/89c13d20-79f6-4a15-9c75-cf199679fc7e-util\") pod \"05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f\" (UID: \"89c13d20-79f6-4a15-9c75-cf199679fc7e\") " pod="openstack-operators/05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f" Dec 06 15:59:03 crc kubenswrapper[4813]: I1206 15:59:03.160892 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/89c13d20-79f6-4a15-9c75-cf199679fc7e-bundle\") pod \"05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f\" (UID: \"89c13d20-79f6-4a15-9c75-cf199679fc7e\") " pod="openstack-operators/05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f" Dec 06 15:59:03 crc kubenswrapper[4813]: I1206 15:59:03.198011 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtsfr\" (UniqueName: \"kubernetes.io/projected/89c13d20-79f6-4a15-9c75-cf199679fc7e-kube-api-access-xtsfr\") pod \"05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f\" (UID: \"89c13d20-79f6-4a15-9c75-cf199679fc7e\") " pod="openstack-operators/05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f" Dec 06 15:59:03 crc kubenswrapper[4813]: I1206 15:59:03.281360 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f" Dec 06 15:59:03 crc kubenswrapper[4813]: I1206 15:59:03.527144 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f"] Dec 06 15:59:03 crc kubenswrapper[4813]: W1206 15:59:03.533393 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89c13d20_79f6_4a15_9c75_cf199679fc7e.slice/crio-cddc1b19453b2a3004a0fe8ec49b320f6ef4f2532441c43dfb76a7afa0c923a3 WatchSource:0}: Error finding container cddc1b19453b2a3004a0fe8ec49b320f6ef4f2532441c43dfb76a7afa0c923a3: Status 404 returned error can't find the container with id cddc1b19453b2a3004a0fe8ec49b320f6ef4f2532441c43dfb76a7afa0c923a3 Dec 06 15:59:03 crc kubenswrapper[4813]: I1206 15:59:03.665154 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f" event={"ID":"89c13d20-79f6-4a15-9c75-cf199679fc7e","Type":"ContainerStarted","Data":"cddc1b19453b2a3004a0fe8ec49b320f6ef4f2532441c43dfb76a7afa0c923a3"} Dec 06 15:59:04 crc kubenswrapper[4813]: I1206 15:59:04.676017 4813 generic.go:334] "Generic (PLEG): container finished" podID="89c13d20-79f6-4a15-9c75-cf199679fc7e" containerID="8a71a27fb2098d4817eeeef1e855bae2e2487aa30a35804930f5a0d6d6a7e598" exitCode=0 Dec 06 15:59:04 crc kubenswrapper[4813]: I1206 15:59:04.676381 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f" event={"ID":"89c13d20-79f6-4a15-9c75-cf199679fc7e","Type":"ContainerDied","Data":"8a71a27fb2098d4817eeeef1e855bae2e2487aa30a35804930f5a0d6d6a7e598"} Dec 06 15:59:05 crc kubenswrapper[4813]: I1206 15:59:05.683576 4813 generic.go:334] "Generic (PLEG): container finished" podID="89c13d20-79f6-4a15-9c75-cf199679fc7e" containerID="ff1e52a39063ca25608933a377a3c797ccf79e0544d44ec19f955ee1d1569408" exitCode=0 Dec 06 15:59:05 crc kubenswrapper[4813]: I1206 15:59:05.683632 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f" event={"ID":"89c13d20-79f6-4a15-9c75-cf199679fc7e","Type":"ContainerDied","Data":"ff1e52a39063ca25608933a377a3c797ccf79e0544d44ec19f955ee1d1569408"} Dec 06 15:59:06 crc kubenswrapper[4813]: I1206 15:59:06.695473 4813 generic.go:334] "Generic (PLEG): container finished" podID="89c13d20-79f6-4a15-9c75-cf199679fc7e" containerID="2f70493bd374da71963bfd9806f923418369611e8c7130707e627d4ab67ab5a5" exitCode=0 Dec 06 15:59:06 crc kubenswrapper[4813]: I1206 15:59:06.695550 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f" event={"ID":"89c13d20-79f6-4a15-9c75-cf199679fc7e","Type":"ContainerDied","Data":"2f70493bd374da71963bfd9806f923418369611e8c7130707e627d4ab67ab5a5"} Dec 06 15:59:08 crc kubenswrapper[4813]: I1206 15:59:08.037786 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f" Dec 06 15:59:08 crc kubenswrapper[4813]: I1206 15:59:08.155587 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/89c13d20-79f6-4a15-9c75-cf199679fc7e-bundle\") pod \"89c13d20-79f6-4a15-9c75-cf199679fc7e\" (UID: \"89c13d20-79f6-4a15-9c75-cf199679fc7e\") " Dec 06 15:59:08 crc kubenswrapper[4813]: I1206 15:59:08.155701 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtsfr\" (UniqueName: \"kubernetes.io/projected/89c13d20-79f6-4a15-9c75-cf199679fc7e-kube-api-access-xtsfr\") pod \"89c13d20-79f6-4a15-9c75-cf199679fc7e\" (UID: \"89c13d20-79f6-4a15-9c75-cf199679fc7e\") " Dec 06 15:59:08 crc kubenswrapper[4813]: I1206 15:59:08.155765 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/89c13d20-79f6-4a15-9c75-cf199679fc7e-util\") pod \"89c13d20-79f6-4a15-9c75-cf199679fc7e\" (UID: \"89c13d20-79f6-4a15-9c75-cf199679fc7e\") " Dec 06 15:59:08 crc kubenswrapper[4813]: I1206 15:59:08.156719 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89c13d20-79f6-4a15-9c75-cf199679fc7e-bundle" (OuterVolumeSpecName: "bundle") pod "89c13d20-79f6-4a15-9c75-cf199679fc7e" (UID: "89c13d20-79f6-4a15-9c75-cf199679fc7e"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:59:08 crc kubenswrapper[4813]: I1206 15:59:08.180479 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89c13d20-79f6-4a15-9c75-cf199679fc7e-kube-api-access-xtsfr" (OuterVolumeSpecName: "kube-api-access-xtsfr") pod "89c13d20-79f6-4a15-9c75-cf199679fc7e" (UID: "89c13d20-79f6-4a15-9c75-cf199679fc7e"). InnerVolumeSpecName "kube-api-access-xtsfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 15:59:08 crc kubenswrapper[4813]: I1206 15:59:08.193042 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89c13d20-79f6-4a15-9c75-cf199679fc7e-util" (OuterVolumeSpecName: "util") pod "89c13d20-79f6-4a15-9c75-cf199679fc7e" (UID: "89c13d20-79f6-4a15-9c75-cf199679fc7e"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 15:59:08 crc kubenswrapper[4813]: I1206 15:59:08.258246 4813 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/89c13d20-79f6-4a15-9c75-cf199679fc7e-util\") on node \"crc\" DevicePath \"\"" Dec 06 15:59:08 crc kubenswrapper[4813]: I1206 15:59:08.258366 4813 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/89c13d20-79f6-4a15-9c75-cf199679fc7e-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 15:59:08 crc kubenswrapper[4813]: I1206 15:59:08.258394 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtsfr\" (UniqueName: \"kubernetes.io/projected/89c13d20-79f6-4a15-9c75-cf199679fc7e-kube-api-access-xtsfr\") on node \"crc\" DevicePath \"\"" Dec 06 15:59:08 crc kubenswrapper[4813]: I1206 15:59:08.713101 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f" event={"ID":"89c13d20-79f6-4a15-9c75-cf199679fc7e","Type":"ContainerDied","Data":"cddc1b19453b2a3004a0fe8ec49b320f6ef4f2532441c43dfb76a7afa0c923a3"} Dec 06 15:59:08 crc kubenswrapper[4813]: I1206 15:59:08.713192 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cddc1b19453b2a3004a0fe8ec49b320f6ef4f2532441c43dfb76a7afa0c923a3" Dec 06 15:59:08 crc kubenswrapper[4813]: I1206 15:59:08.713167 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f" Dec 06 15:59:15 crc kubenswrapper[4813]: I1206 15:59:15.127231 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-cb74cdcd8-r5tfz"] Dec 06 15:59:15 crc kubenswrapper[4813]: E1206 15:59:15.127672 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89c13d20-79f6-4a15-9c75-cf199679fc7e" containerName="extract" Dec 06 15:59:15 crc kubenswrapper[4813]: I1206 15:59:15.127684 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="89c13d20-79f6-4a15-9c75-cf199679fc7e" containerName="extract" Dec 06 15:59:15 crc kubenswrapper[4813]: E1206 15:59:15.127698 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89c13d20-79f6-4a15-9c75-cf199679fc7e" containerName="pull" Dec 06 15:59:15 crc kubenswrapper[4813]: I1206 15:59:15.127704 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="89c13d20-79f6-4a15-9c75-cf199679fc7e" containerName="pull" Dec 06 15:59:15 crc kubenswrapper[4813]: E1206 15:59:15.127718 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89c13d20-79f6-4a15-9c75-cf199679fc7e" containerName="util" Dec 06 15:59:15 crc kubenswrapper[4813]: I1206 15:59:15.127724 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="89c13d20-79f6-4a15-9c75-cf199679fc7e" containerName="util" Dec 06 15:59:15 crc kubenswrapper[4813]: I1206 15:59:15.127820 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="89c13d20-79f6-4a15-9c75-cf199679fc7e" containerName="extract" Dec 06 15:59:15 crc kubenswrapper[4813]: I1206 15:59:15.128178 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-cb74cdcd8-r5tfz" Dec 06 15:59:15 crc kubenswrapper[4813]: I1206 15:59:15.130828 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-9x4gd" Dec 06 15:59:15 crc kubenswrapper[4813]: I1206 15:59:15.169309 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-cb74cdcd8-r5tfz"] Dec 06 15:59:15 crc kubenswrapper[4813]: I1206 15:59:15.253123 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bz7pc\" (UniqueName: \"kubernetes.io/projected/d2bd6dc9-ba04-4168-b879-415d9ce263d4-kube-api-access-bz7pc\") pod \"openstack-operator-controller-operator-cb74cdcd8-r5tfz\" (UID: \"d2bd6dc9-ba04-4168-b879-415d9ce263d4\") " pod="openstack-operators/openstack-operator-controller-operator-cb74cdcd8-r5tfz" Dec 06 15:59:15 crc kubenswrapper[4813]: I1206 15:59:15.354395 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bz7pc\" (UniqueName: \"kubernetes.io/projected/d2bd6dc9-ba04-4168-b879-415d9ce263d4-kube-api-access-bz7pc\") pod \"openstack-operator-controller-operator-cb74cdcd8-r5tfz\" (UID: \"d2bd6dc9-ba04-4168-b879-415d9ce263d4\") " pod="openstack-operators/openstack-operator-controller-operator-cb74cdcd8-r5tfz" Dec 06 15:59:15 crc kubenswrapper[4813]: I1206 15:59:15.382846 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bz7pc\" (UniqueName: \"kubernetes.io/projected/d2bd6dc9-ba04-4168-b879-415d9ce263d4-kube-api-access-bz7pc\") pod \"openstack-operator-controller-operator-cb74cdcd8-r5tfz\" (UID: \"d2bd6dc9-ba04-4168-b879-415d9ce263d4\") " pod="openstack-operators/openstack-operator-controller-operator-cb74cdcd8-r5tfz" Dec 06 15:59:15 crc kubenswrapper[4813]: I1206 15:59:15.443688 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-cb74cdcd8-r5tfz" Dec 06 15:59:15 crc kubenswrapper[4813]: I1206 15:59:15.746911 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-cb74cdcd8-r5tfz"] Dec 06 15:59:15 crc kubenswrapper[4813]: I1206 15:59:15.762055 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-cb74cdcd8-r5tfz" event={"ID":"d2bd6dc9-ba04-4168-b879-415d9ce263d4","Type":"ContainerStarted","Data":"62c386b0fa97c3caa353a3c9bb26f6982e331ad69070c416536b1ce46b86478c"} Dec 06 15:59:19 crc kubenswrapper[4813]: I1206 15:59:19.434789 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 15:59:19 crc kubenswrapper[4813]: I1206 15:59:19.435180 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 15:59:20 crc kubenswrapper[4813]: I1206 15:59:20.789190 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-cb74cdcd8-r5tfz" event={"ID":"d2bd6dc9-ba04-4168-b879-415d9ce263d4","Type":"ContainerStarted","Data":"5891ffcd0b7066aac5403c0cf2fbc05443d8cad5eab72dc3892165b34ec727dc"} Dec 06 15:59:20 crc kubenswrapper[4813]: I1206 15:59:20.790330 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-cb74cdcd8-r5tfz" Dec 06 15:59:20 crc kubenswrapper[4813]: I1206 15:59:20.821890 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-cb74cdcd8-r5tfz" podStartSLOduration=1.098459298 podStartE2EDuration="5.821867014s" podCreationTimestamp="2025-12-06 15:59:15 +0000 UTC" firstStartedPulling="2025-12-06 15:59:15.748000343 +0000 UTC m=+795.638879919" lastFinishedPulling="2025-12-06 15:59:20.471408049 +0000 UTC m=+800.362287635" observedRunningTime="2025-12-06 15:59:20.818928776 +0000 UTC m=+800.709808382" watchObservedRunningTime="2025-12-06 15:59:20.821867014 +0000 UTC m=+800.712746630" Dec 06 15:59:25 crc kubenswrapper[4813]: I1206 15:59:25.448342 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-cb74cdcd8-r5tfz" Dec 06 15:59:44 crc kubenswrapper[4813]: I1206 15:59:44.773524 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6dc7dc95b4-6wznz"] Dec 06 15:59:44 crc kubenswrapper[4813]: I1206 15:59:44.774740 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6dc7dc95b4-6wznz" Dec 06 15:59:44 crc kubenswrapper[4813]: I1206 15:59:44.779987 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-5dlpj" Dec 06 15:59:44 crc kubenswrapper[4813]: I1206 15:59:44.784313 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-8mt55"] Dec 06 15:59:44 crc kubenswrapper[4813]: I1206 15:59:44.785227 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8mt55" Dec 06 15:59:44 crc kubenswrapper[4813]: I1206 15:59:44.786247 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-ccvtg" Dec 06 15:59:44 crc kubenswrapper[4813]: I1206 15:59:44.812943 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-4rdzx"] Dec 06 15:59:44 crc kubenswrapper[4813]: I1206 15:59:44.817539 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-4rdzx" Dec 06 15:59:44 crc kubenswrapper[4813]: I1206 15:59:44.830890 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-cq78w" Dec 06 15:59:44 crc kubenswrapper[4813]: I1206 15:59:44.864384 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-8mt55"] Dec 06 15:59:44 crc kubenswrapper[4813]: I1206 15:59:44.870250 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-4rdzx"] Dec 06 15:59:44 crc kubenswrapper[4813]: I1206 15:59:44.899000 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-f9nfd"] Dec 06 15:59:44 crc kubenswrapper[4813]: I1206 15:59:44.900019 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-f9nfd" Dec 06 15:59:44 crc kubenswrapper[4813]: I1206 15:59:44.901056 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffndd\" (UniqueName: \"kubernetes.io/projected/218be173-f41f-48b1-8af0-b94bbc483e8b-kube-api-access-ffndd\") pod \"designate-operator-controller-manager-697fb699cf-4rdzx\" (UID: \"218be173-f41f-48b1-8af0-b94bbc483e8b\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-4rdzx" Dec 06 15:59:44 crc kubenswrapper[4813]: I1206 15:59:44.901111 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9mpn\" (UniqueName: \"kubernetes.io/projected/a8b0c607-9997-44e8-a3cc-a36e4b2e70d2-kube-api-access-h9mpn\") pod \"cinder-operator-controller-manager-6c677c69b-8mt55\" (UID: \"a8b0c607-9997-44e8-a3cc-a36e4b2e70d2\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8mt55" Dec 06 15:59:44 crc kubenswrapper[4813]: I1206 15:59:44.901131 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjq65\" (UniqueName: \"kubernetes.io/projected/24d6dd0f-86e6-4b7a-beeb-56025a76ab5a-kube-api-access-fjq65\") pod \"barbican-operator-controller-manager-6dc7dc95b4-6wznz\" (UID: \"24d6dd0f-86e6-4b7a-beeb-56025a76ab5a\") " pod="openstack-operators/barbican-operator-controller-manager-6dc7dc95b4-6wznz" Dec 06 15:59:44 crc kubenswrapper[4813]: I1206 15:59:44.907865 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-szzxr" Dec 06 15:59:44 crc kubenswrapper[4813]: I1206 15:59:44.914126 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6dc7dc95b4-6wznz"] Dec 06 15:59:44 crc kubenswrapper[4813]: I1206 15:59:44.930079 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-wxsqx"] Dec 06 15:59:44 crc kubenswrapper[4813]: I1206 15:59:44.941376 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-wxsqx" Dec 06 15:59:44 crc kubenswrapper[4813]: I1206 15:59:44.948803 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-4pjn9" Dec 06 15:59:44 crc kubenswrapper[4813]: I1206 15:59:44.960219 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-f9nfd"] Dec 06 15:59:44 crc kubenswrapper[4813]: I1206 15:59:44.992618 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-wxsqx"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.002974 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-g4ztj"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.003905 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-g4ztj" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.003935 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffndd\" (UniqueName: \"kubernetes.io/projected/218be173-f41f-48b1-8af0-b94bbc483e8b-kube-api-access-ffndd\") pod \"designate-operator-controller-manager-697fb699cf-4rdzx\" (UID: \"218be173-f41f-48b1-8af0-b94bbc483e8b\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-4rdzx" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.004006 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9mpn\" (UniqueName: \"kubernetes.io/projected/a8b0c607-9997-44e8-a3cc-a36e4b2e70d2-kube-api-access-h9mpn\") pod \"cinder-operator-controller-manager-6c677c69b-8mt55\" (UID: \"a8b0c607-9997-44e8-a3cc-a36e4b2e70d2\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8mt55" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.004123 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6x57\" (UniqueName: \"kubernetes.io/projected/3edd7ca8-a3ba-46c5-96c8-77aa9be672b7-kube-api-access-x6x57\") pod \"glance-operator-controller-manager-5697bb5779-f9nfd\" (UID: \"3edd7ca8-a3ba-46c5-96c8-77aa9be672b7\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-f9nfd" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.004143 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjq65\" (UniqueName: \"kubernetes.io/projected/24d6dd0f-86e6-4b7a-beeb-56025a76ab5a-kube-api-access-fjq65\") pod \"barbican-operator-controller-manager-6dc7dc95b4-6wznz\" (UID: \"24d6dd0f-86e6-4b7a-beeb-56025a76ab5a\") " pod="openstack-operators/barbican-operator-controller-manager-6dc7dc95b4-6wznz" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.004295 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7p57p\" (UniqueName: \"kubernetes.io/projected/e33ae276-0e3d-44b5-b970-17e98051c49c-kube-api-access-7p57p\") pod \"heat-operator-controller-manager-5f64f6f8bb-wxsqx\" (UID: \"e33ae276-0e3d-44b5-b970-17e98051c49c\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-wxsqx" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.013661 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-jsn79" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.023173 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-pkvxj"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.024101 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pkvxj" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.028695 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.029026 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-l9rd4" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.035564 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-g4ztj"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.037605 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjq65\" (UniqueName: \"kubernetes.io/projected/24d6dd0f-86e6-4b7a-beeb-56025a76ab5a-kube-api-access-fjq65\") pod \"barbican-operator-controller-manager-6dc7dc95b4-6wznz\" (UID: \"24d6dd0f-86e6-4b7a-beeb-56025a76ab5a\") " pod="openstack-operators/barbican-operator-controller-manager-6dc7dc95b4-6wznz" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.046928 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9mpn\" (UniqueName: \"kubernetes.io/projected/a8b0c607-9997-44e8-a3cc-a36e4b2e70d2-kube-api-access-h9mpn\") pod \"cinder-operator-controller-manager-6c677c69b-8mt55\" (UID: \"a8b0c607-9997-44e8-a3cc-a36e4b2e70d2\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8mt55" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.059840 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffndd\" (UniqueName: \"kubernetes.io/projected/218be173-f41f-48b1-8af0-b94bbc483e8b-kube-api-access-ffndd\") pod \"designate-operator-controller-manager-697fb699cf-4rdzx\" (UID: \"218be173-f41f-48b1-8af0-b94bbc483e8b\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-4rdzx" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.071650 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-cxh72"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.073122 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-cxh72" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.075056 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-tbhk9" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.096581 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6dc7dc95b4-6wznz" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.107286 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7p57p\" (UniqueName: \"kubernetes.io/projected/e33ae276-0e3d-44b5-b970-17e98051c49c-kube-api-access-7p57p\") pod \"heat-operator-controller-manager-5f64f6f8bb-wxsqx\" (UID: \"e33ae276-0e3d-44b5-b970-17e98051c49c\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-wxsqx" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.107332 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zptcz\" (UniqueName: \"kubernetes.io/projected/d67349e5-13fb-4859-ac8d-3e2c6bd67dbb-kube-api-access-zptcz\") pod \"horizon-operator-controller-manager-68c6d99b8f-g4ztj\" (UID: \"d67349e5-13fb-4859-ac8d-3e2c6bd67dbb\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-g4ztj" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.107367 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxzdf\" (UniqueName: \"kubernetes.io/projected/6983648f-60ef-429d-a2fd-de5dc7e6d0ba-kube-api-access-cxzdf\") pod \"infra-operator-controller-manager-78d48bff9d-pkvxj\" (UID: \"6983648f-60ef-429d-a2fd-de5dc7e6d0ba\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pkvxj" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.107386 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6983648f-60ef-429d-a2fd-de5dc7e6d0ba-cert\") pod \"infra-operator-controller-manager-78d48bff9d-pkvxj\" (UID: \"6983648f-60ef-429d-a2fd-de5dc7e6d0ba\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pkvxj" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.107433 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6x57\" (UniqueName: \"kubernetes.io/projected/3edd7ca8-a3ba-46c5-96c8-77aa9be672b7-kube-api-access-x6x57\") pod \"glance-operator-controller-manager-5697bb5779-f9nfd\" (UID: \"3edd7ca8-a3ba-46c5-96c8-77aa9be672b7\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-f9nfd" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.115439 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-cxh72"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.127358 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-pkvxj"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.128210 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-cjzht"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.129128 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-cjzht" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.131089 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8mt55" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.146700 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-7l766" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.147141 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-cjzht"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.161194 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-4rdzx" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.206864 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6x57\" (UniqueName: \"kubernetes.io/projected/3edd7ca8-a3ba-46c5-96c8-77aa9be672b7-kube-api-access-x6x57\") pod \"glance-operator-controller-manager-5697bb5779-f9nfd\" (UID: \"3edd7ca8-a3ba-46c5-96c8-77aa9be672b7\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-f9nfd" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.209083 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zptcz\" (UniqueName: \"kubernetes.io/projected/d67349e5-13fb-4859-ac8d-3e2c6bd67dbb-kube-api-access-zptcz\") pod \"horizon-operator-controller-manager-68c6d99b8f-g4ztj\" (UID: \"d67349e5-13fb-4859-ac8d-3e2c6bd67dbb\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-g4ztj" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.209137 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxzdf\" (UniqueName: \"kubernetes.io/projected/6983648f-60ef-429d-a2fd-de5dc7e6d0ba-kube-api-access-cxzdf\") pod \"infra-operator-controller-manager-78d48bff9d-pkvxj\" (UID: \"6983648f-60ef-429d-a2fd-de5dc7e6d0ba\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pkvxj" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.209160 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6983648f-60ef-429d-a2fd-de5dc7e6d0ba-cert\") pod \"infra-operator-controller-manager-78d48bff9d-pkvxj\" (UID: \"6983648f-60ef-429d-a2fd-de5dc7e6d0ba\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pkvxj" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.209220 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xrkh\" (UniqueName: \"kubernetes.io/projected/4d2f63b5-c93e-4949-bb79-459dad74dcf8-kube-api-access-9xrkh\") pod \"ironic-operator-controller-manager-967d97867-cxh72\" (UID: \"4d2f63b5-c93e-4949-bb79-459dad74dcf8\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-cxh72" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.209238 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knwhm\" (UniqueName: \"kubernetes.io/projected/8870fd51-2141-4cc8-956b-050e849dcbd4-kube-api-access-knwhm\") pod \"keystone-operator-controller-manager-7765d96ddf-cjzht\" (UID: \"8870fd51-2141-4cc8-956b-050e849dcbd4\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-cjzht" Dec 06 15:59:45 crc kubenswrapper[4813]: E1206 15:59:45.209471 4813 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 15:59:45 crc kubenswrapper[4813]: E1206 15:59:45.209526 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6983648f-60ef-429d-a2fd-de5dc7e6d0ba-cert podName:6983648f-60ef-429d-a2fd-de5dc7e6d0ba nodeName:}" failed. No retries permitted until 2025-12-06 15:59:45.709508918 +0000 UTC m=+825.600388494 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6983648f-60ef-429d-a2fd-de5dc7e6d0ba-cert") pod "infra-operator-controller-manager-78d48bff9d-pkvxj" (UID: "6983648f-60ef-429d-a2fd-de5dc7e6d0ba") : secret "infra-operator-webhook-server-cert" not found Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.220842 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7p57p\" (UniqueName: \"kubernetes.io/projected/e33ae276-0e3d-44b5-b970-17e98051c49c-kube-api-access-7p57p\") pod \"heat-operator-controller-manager-5f64f6f8bb-wxsqx\" (UID: \"e33ae276-0e3d-44b5-b970-17e98051c49c\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-wxsqx" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.243046 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-f9nfd" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.243360 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6n2hj"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.244441 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6n2hj" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.248110 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-kjp8h" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.260495 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-6p6sf"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.261726 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-6p6sf" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.265309 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxzdf\" (UniqueName: \"kubernetes.io/projected/6983648f-60ef-429d-a2fd-de5dc7e6d0ba-kube-api-access-cxzdf\") pod \"infra-operator-controller-manager-78d48bff9d-pkvxj\" (UID: \"6983648f-60ef-429d-a2fd-de5dc7e6d0ba\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pkvxj" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.277399 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-6rr2s"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.278542 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-6rr2s" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.283891 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-gp2lg" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.284289 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-gcwfq" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.294075 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-wxsqx" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.296178 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zptcz\" (UniqueName: \"kubernetes.io/projected/d67349e5-13fb-4859-ac8d-3e2c6bd67dbb-kube-api-access-zptcz\") pod \"horizon-operator-controller-manager-68c6d99b8f-g4ztj\" (UID: \"d67349e5-13fb-4859-ac8d-3e2c6bd67dbb\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-g4ztj" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.298838 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-6rr2s"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.309887 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xrkh\" (UniqueName: \"kubernetes.io/projected/4d2f63b5-c93e-4949-bb79-459dad74dcf8-kube-api-access-9xrkh\") pod \"ironic-operator-controller-manager-967d97867-cxh72\" (UID: \"4d2f63b5-c93e-4949-bb79-459dad74dcf8\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-cxh72" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.309924 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knwhm\" (UniqueName: \"kubernetes.io/projected/8870fd51-2141-4cc8-956b-050e849dcbd4-kube-api-access-knwhm\") pod \"keystone-operator-controller-manager-7765d96ddf-cjzht\" (UID: \"8870fd51-2141-4cc8-956b-050e849dcbd4\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-cjzht" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.309953 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7z7fw\" (UniqueName: \"kubernetes.io/projected/99a0eae7-c990-42cd-a18e-78b81177bbb6-kube-api-access-7z7fw\") pod \"mariadb-operator-controller-manager-79c8c4686c-6n2hj\" (UID: \"99a0eae7-c990-42cd-a18e-78b81177bbb6\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6n2hj" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.309982 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cccx4\" (UniqueName: \"kubernetes.io/projected/a8763d41-9404-43ec-866b-9244bf2c4af9-kube-api-access-cccx4\") pod \"manila-operator-controller-manager-5b5fd79c9c-6p6sf\" (UID: \"a8763d41-9404-43ec-866b-9244bf2c4af9\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-6p6sf" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.310005 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2cv9\" (UniqueName: \"kubernetes.io/projected/b4cd1114-c317-4351-b566-05317dd589f1-kube-api-access-t2cv9\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-6rr2s\" (UID: \"b4cd1114-c317-4351-b566-05317dd589f1\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-6rr2s" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.335621 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-g4ztj" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.342485 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-78fv6"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.343641 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-78fv6" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.351635 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-6p6sf"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.354020 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xrkh\" (UniqueName: \"kubernetes.io/projected/4d2f63b5-c93e-4949-bb79-459dad74dcf8-kube-api-access-9xrkh\") pod \"ironic-operator-controller-manager-967d97867-cxh72\" (UID: \"4d2f63b5-c93e-4949-bb79-459dad74dcf8\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-cxh72" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.354029 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-cxnzw" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.364333 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6n2hj"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.380074 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-78fv6"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.381554 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knwhm\" (UniqueName: \"kubernetes.io/projected/8870fd51-2141-4cc8-956b-050e849dcbd4-kube-api-access-knwhm\") pod \"keystone-operator-controller-manager-7765d96ddf-cjzht\" (UID: \"8870fd51-2141-4cc8-956b-050e849dcbd4\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-cjzht" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.388455 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-g9q96"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.389894 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-g9q96" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.402088 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-k2lfd" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.411503 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7z7fw\" (UniqueName: \"kubernetes.io/projected/99a0eae7-c990-42cd-a18e-78b81177bbb6-kube-api-access-7z7fw\") pod \"mariadb-operator-controller-manager-79c8c4686c-6n2hj\" (UID: \"99a0eae7-c990-42cd-a18e-78b81177bbb6\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6n2hj" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.411547 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cccx4\" (UniqueName: \"kubernetes.io/projected/a8763d41-9404-43ec-866b-9244bf2c4af9-kube-api-access-cccx4\") pod \"manila-operator-controller-manager-5b5fd79c9c-6p6sf\" (UID: \"a8763d41-9404-43ec-866b-9244bf2c4af9\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-6p6sf" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.411574 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2cv9\" (UniqueName: \"kubernetes.io/projected/b4cd1114-c317-4351-b566-05317dd589f1-kube-api-access-t2cv9\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-6rr2s\" (UID: \"b4cd1114-c317-4351-b566-05317dd589f1\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-6rr2s" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.411638 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcdhd\" (UniqueName: \"kubernetes.io/projected/b675acf0-51cc-43d9-b9c6-bd0ee1730d25-kube-api-access-hcdhd\") pod \"nova-operator-controller-manager-697bc559fc-78fv6\" (UID: \"b675acf0-51cc-43d9-b9c6-bd0ee1730d25\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-78fv6" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.425361 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-cxh72" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.457876 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-g9q96"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.460463 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fc7jrk"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.464974 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2cv9\" (UniqueName: \"kubernetes.io/projected/b4cd1114-c317-4351-b566-05317dd589f1-kube-api-access-t2cv9\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-6rr2s\" (UID: \"b4cd1114-c317-4351-b566-05317dd589f1\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-6rr2s" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.465788 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7z7fw\" (UniqueName: \"kubernetes.io/projected/99a0eae7-c990-42cd-a18e-78b81177bbb6-kube-api-access-7z7fw\") pod \"mariadb-operator-controller-manager-79c8c4686c-6n2hj\" (UID: \"99a0eae7-c990-42cd-a18e-78b81177bbb6\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6n2hj" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.506744 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-cjzht" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.510058 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fc7jrk" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.515784 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-rrd7f" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.515963 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.517966 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ndhf\" (UniqueName: \"kubernetes.io/projected/031a9bd1-298a-4945-a52d-452005c8467e-kube-api-access-6ndhf\") pod \"octavia-operator-controller-manager-998648c74-g9q96\" (UID: \"031a9bd1-298a-4945-a52d-452005c8467e\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-g9q96" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.518032 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcdhd\" (UniqueName: \"kubernetes.io/projected/b675acf0-51cc-43d9-b9c6-bd0ee1730d25-kube-api-access-hcdhd\") pod \"nova-operator-controller-manager-697bc559fc-78fv6\" (UID: \"b675acf0-51cc-43d9-b9c6-bd0ee1730d25\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-78fv6" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.523530 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cccx4\" (UniqueName: \"kubernetes.io/projected/a8763d41-9404-43ec-866b-9244bf2c4af9-kube-api-access-cccx4\") pod \"manila-operator-controller-manager-5b5fd79c9c-6p6sf\" (UID: \"a8763d41-9404-43ec-866b-9244bf2c4af9\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-6p6sf" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.544220 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-2mpwn"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.565147 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-2mpwn" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.576451 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-6kzjc" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.590238 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6n2hj" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.615685 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcdhd\" (UniqueName: \"kubernetes.io/projected/b675acf0-51cc-43d9-b9c6-bd0ee1730d25-kube-api-access-hcdhd\") pod \"nova-operator-controller-manager-697bc559fc-78fv6\" (UID: \"b675acf0-51cc-43d9-b9c6-bd0ee1730d25\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-78fv6" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.622642 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ndhf\" (UniqueName: \"kubernetes.io/projected/031a9bd1-298a-4945-a52d-452005c8467e-kube-api-access-6ndhf\") pod \"octavia-operator-controller-manager-998648c74-g9q96\" (UID: \"031a9bd1-298a-4945-a52d-452005c8467e\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-g9q96" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.637748 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-6p6sf" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.644554 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fc7jrk"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.645553 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-6rr2s" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.663328 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-rsbqf"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.664489 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-rsbqf" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.667794 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-zzfxm" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.689683 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-78fv6" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.692332 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ndhf\" (UniqueName: \"kubernetes.io/projected/031a9bd1-298a-4945-a52d-452005c8467e-kube-api-access-6ndhf\") pod \"octavia-operator-controller-manager-998648c74-g9q96\" (UID: \"031a9bd1-298a-4945-a52d-452005c8467e\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-g9q96" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.697812 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-x8vkl"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.698968 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-x8vkl" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.703954 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-d6pfx" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.704243 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-2mpwn"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.713422 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-rsbqf"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.713682 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-g9q96" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.717185 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-x8vkl"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.726354 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vczch\" (UniqueName: \"kubernetes.io/projected/279a2707-03fd-4cbc-9e13-d7a0d2a284db-kube-api-access-vczch\") pod \"openstack-baremetal-operator-controller-manager-84b575879fc7jrk\" (UID: \"279a2707-03fd-4cbc-9e13-d7a0d2a284db\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fc7jrk" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.726735 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/279a2707-03fd-4cbc-9e13-d7a0d2a284db-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fc7jrk\" (UID: \"279a2707-03fd-4cbc-9e13-d7a0d2a284db\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fc7jrk" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.727179 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6983648f-60ef-429d-a2fd-de5dc7e6d0ba-cert\") pod \"infra-operator-controller-manager-78d48bff9d-pkvxj\" (UID: \"6983648f-60ef-429d-a2fd-de5dc7e6d0ba\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pkvxj" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.727405 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbdgh\" (UniqueName: \"kubernetes.io/projected/c35efae6-9bf9-466d-ac8d-203d8a2c4b97-kube-api-access-xbdgh\") pod \"ovn-operator-controller-manager-b6456fdb6-2mpwn\" (UID: \"c35efae6-9bf9-466d-ac8d-203d8a2c4b97\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-2mpwn" Dec 06 15:59:45 crc kubenswrapper[4813]: E1206 15:59:45.727540 4813 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 15:59:45 crc kubenswrapper[4813]: E1206 15:59:45.727597 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6983648f-60ef-429d-a2fd-de5dc7e6d0ba-cert podName:6983648f-60ef-429d-a2fd-de5dc7e6d0ba nodeName:}" failed. No retries permitted until 2025-12-06 15:59:46.727579419 +0000 UTC m=+826.618458995 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6983648f-60ef-429d-a2fd-de5dc7e6d0ba-cert") pod "infra-operator-controller-manager-78d48bff9d-pkvxj" (UID: "6983648f-60ef-429d-a2fd-de5dc7e6d0ba") : secret "infra-operator-webhook-server-cert" not found Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.727863 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qkrqw"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.729079 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qkrqw" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.732576 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-6df42" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.739567 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-cbmwf"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.740763 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-cbmwf" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.765885 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-7qxks" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.779327 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qkrqw"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.830793 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/279a2707-03fd-4cbc-9e13-d7a0d2a284db-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fc7jrk\" (UID: \"279a2707-03fd-4cbc-9e13-d7a0d2a284db\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fc7jrk" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.830890 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knrft\" (UniqueName: \"kubernetes.io/projected/67c6b7ea-97cf-4ffa-9f3b-5111d4fbdeba-kube-api-access-knrft\") pod \"placement-operator-controller-manager-78f8948974-rsbqf\" (UID: \"67c6b7ea-97cf-4ffa-9f3b-5111d4fbdeba\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-rsbqf" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.830910 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbdgh\" (UniqueName: \"kubernetes.io/projected/c35efae6-9bf9-466d-ac8d-203d8a2c4b97-kube-api-access-xbdgh\") pod \"ovn-operator-controller-manager-b6456fdb6-2mpwn\" (UID: \"c35efae6-9bf9-466d-ac8d-203d8a2c4b97\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-2mpwn" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.830935 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmt5c\" (UniqueName: \"kubernetes.io/projected/6d54c577-3fdf-48af-9070-301dd3b3eb81-kube-api-access-kmt5c\") pod \"telemetry-operator-controller-manager-58d5ff84df-qkrqw\" (UID: \"6d54c577-3fdf-48af-9070-301dd3b3eb81\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qkrqw" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.830994 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vczch\" (UniqueName: \"kubernetes.io/projected/279a2707-03fd-4cbc-9e13-d7a0d2a284db-kube-api-access-vczch\") pod \"openstack-baremetal-operator-controller-manager-84b575879fc7jrk\" (UID: \"279a2707-03fd-4cbc-9e13-d7a0d2a284db\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fc7jrk" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.831017 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwj4p\" (UniqueName: \"kubernetes.io/projected/b05dafaa-0d81-4b80-ad3e-48da48a1655f-kube-api-access-bwj4p\") pod \"swift-operator-controller-manager-9d58d64bc-x8vkl\" (UID: \"b05dafaa-0d81-4b80-ad3e-48da48a1655f\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-x8vkl" Dec 06 15:59:45 crc kubenswrapper[4813]: E1206 15:59:45.830789 4813 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 15:59:45 crc kubenswrapper[4813]: E1206 15:59:45.831158 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/279a2707-03fd-4cbc-9e13-d7a0d2a284db-cert podName:279a2707-03fd-4cbc-9e13-d7a0d2a284db nodeName:}" failed. No retries permitted until 2025-12-06 15:59:46.331126733 +0000 UTC m=+826.222006309 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/279a2707-03fd-4cbc-9e13-d7a0d2a284db-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fc7jrk" (UID: "279a2707-03fd-4cbc-9e13-d7a0d2a284db") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.884996 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbdgh\" (UniqueName: \"kubernetes.io/projected/c35efae6-9bf9-466d-ac8d-203d8a2c4b97-kube-api-access-xbdgh\") pod \"ovn-operator-controller-manager-b6456fdb6-2mpwn\" (UID: \"c35efae6-9bf9-466d-ac8d-203d8a2c4b97\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-2mpwn" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.885455 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-cbmwf"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.894326 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vczch\" (UniqueName: \"kubernetes.io/projected/279a2707-03fd-4cbc-9e13-d7a0d2a284db-kube-api-access-vczch\") pod \"openstack-baremetal-operator-controller-manager-84b575879fc7jrk\" (UID: \"279a2707-03fd-4cbc-9e13-d7a0d2a284db\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fc7jrk" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.923811 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-2mpwn" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.934651 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwj4p\" (UniqueName: \"kubernetes.io/projected/b05dafaa-0d81-4b80-ad3e-48da48a1655f-kube-api-access-bwj4p\") pod \"swift-operator-controller-manager-9d58d64bc-x8vkl\" (UID: \"b05dafaa-0d81-4b80-ad3e-48da48a1655f\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-x8vkl" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.934719 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pn8x\" (UniqueName: \"kubernetes.io/projected/28acd95a-e657-46b5-a771-a24b27bdc6fe-kube-api-access-9pn8x\") pod \"test-operator-controller-manager-5854674fcc-cbmwf\" (UID: \"28acd95a-e657-46b5-a771-a24b27bdc6fe\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-cbmwf" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.934769 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knrft\" (UniqueName: \"kubernetes.io/projected/67c6b7ea-97cf-4ffa-9f3b-5111d4fbdeba-kube-api-access-knrft\") pod \"placement-operator-controller-manager-78f8948974-rsbqf\" (UID: \"67c6b7ea-97cf-4ffa-9f3b-5111d4fbdeba\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-rsbqf" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.934798 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmt5c\" (UniqueName: \"kubernetes.io/projected/6d54c577-3fdf-48af-9070-301dd3b3eb81-kube-api-access-kmt5c\") pod \"telemetry-operator-controller-manager-58d5ff84df-qkrqw\" (UID: \"6d54c577-3fdf-48af-9070-301dd3b3eb81\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qkrqw" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.939066 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-8zvf5"] Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.964776 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmt5c\" (UniqueName: \"kubernetes.io/projected/6d54c577-3fdf-48af-9070-301dd3b3eb81-kube-api-access-kmt5c\") pod \"telemetry-operator-controller-manager-58d5ff84df-qkrqw\" (UID: \"6d54c577-3fdf-48af-9070-301dd3b3eb81\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qkrqw" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.972106 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knrft\" (UniqueName: \"kubernetes.io/projected/67c6b7ea-97cf-4ffa-9f3b-5111d4fbdeba-kube-api-access-knrft\") pod \"placement-operator-controller-manager-78f8948974-rsbqf\" (UID: \"67c6b7ea-97cf-4ffa-9f3b-5111d4fbdeba\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-rsbqf" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.978538 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-8zvf5" Dec 06 15:59:45 crc kubenswrapper[4813]: I1206 15:59:45.981390 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-2gzqs" Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.018158 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwj4p\" (UniqueName: \"kubernetes.io/projected/b05dafaa-0d81-4b80-ad3e-48da48a1655f-kube-api-access-bwj4p\") pod \"swift-operator-controller-manager-9d58d64bc-x8vkl\" (UID: \"b05dafaa-0d81-4b80-ad3e-48da48a1655f\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-x8vkl" Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.021431 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-8zvf5"] Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.036072 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pn8x\" (UniqueName: \"kubernetes.io/projected/28acd95a-e657-46b5-a771-a24b27bdc6fe-kube-api-access-9pn8x\") pod \"test-operator-controller-manager-5854674fcc-cbmwf\" (UID: \"28acd95a-e657-46b5-a771-a24b27bdc6fe\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-cbmwf" Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.066502 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-rsbqf" Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.068279 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pn8x\" (UniqueName: \"kubernetes.io/projected/28acd95a-e657-46b5-a771-a24b27bdc6fe-kube-api-access-9pn8x\") pod \"test-operator-controller-manager-5854674fcc-cbmwf\" (UID: \"28acd95a-e657-46b5-a771-a24b27bdc6fe\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-cbmwf" Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.083328 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26"] Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.084222 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26" Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.097110 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-x8vkl" Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.098166 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.098396 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 06 15:59:46 crc kubenswrapper[4813]: W1206 15:59:46.098447 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod218be173_f41f_48b1_8af0_b94bbc483e8b.slice/crio-0b2b439c624d2af893b660101890e402ffa88120103e0f89de86b47076f0edf0 WatchSource:0}: Error finding container 0b2b439c624d2af893b660101890e402ffa88120103e0f89de86b47076f0edf0: Status 404 returned error can't find the container with id 0b2b439c624d2af893b660101890e402ffa88120103e0f89de86b47076f0edf0 Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.098536 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-t6rxr" Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.103370 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26"] Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.136357 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qkrqw" Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.146198 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qswn\" (UniqueName: \"kubernetes.io/projected/5a8652b7-8a55-4658-acce-b7573421b94d-kube-api-access-8qswn\") pod \"watcher-operator-controller-manager-667bd8d554-8zvf5\" (UID: \"5a8652b7-8a55-4658-acce-b7573421b94d\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-8zvf5" Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.168644 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9lwhh"] Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.169498 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9lwhh" Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.187925 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-cbmwf" Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.188549 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-nvc4g" Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.222435 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9lwhh"] Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.247197 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-webhook-certs\") pod \"openstack-operator-controller-manager-5c8695d769-78g26\" (UID: \"572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b\") " pod="openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26" Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.247254 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qswn\" (UniqueName: \"kubernetes.io/projected/5a8652b7-8a55-4658-acce-b7573421b94d-kube-api-access-8qswn\") pod \"watcher-operator-controller-manager-667bd8d554-8zvf5\" (UID: \"5a8652b7-8a55-4658-acce-b7573421b94d\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-8zvf5" Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.247331 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-metrics-certs\") pod \"openstack-operator-controller-manager-5c8695d769-78g26\" (UID: \"572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b\") " pod="openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26" Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.247356 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98kvs\" (UniqueName: \"kubernetes.io/projected/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-kube-api-access-98kvs\") pod \"openstack-operator-controller-manager-5c8695d769-78g26\" (UID: \"572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b\") " pod="openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26" Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.264325 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-4rdzx"] Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.273852 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qswn\" (UniqueName: \"kubernetes.io/projected/5a8652b7-8a55-4658-acce-b7573421b94d-kube-api-access-8qswn\") pod \"watcher-operator-controller-manager-667bd8d554-8zvf5\" (UID: \"5a8652b7-8a55-4658-acce-b7573421b94d\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-8zvf5" Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.312575 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-8zvf5" Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.348595 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/279a2707-03fd-4cbc-9e13-d7a0d2a284db-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fc7jrk\" (UID: \"279a2707-03fd-4cbc-9e13-d7a0d2a284db\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fc7jrk" Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.348639 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqc5x\" (UniqueName: \"kubernetes.io/projected/00cfc20a-68d6-46be-98b8-83a43c9d11d9-kube-api-access-rqc5x\") pod \"rabbitmq-cluster-operator-manager-668c99d594-9lwhh\" (UID: \"00cfc20a-68d6-46be-98b8-83a43c9d11d9\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9lwhh" Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.348690 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-webhook-certs\") pod \"openstack-operator-controller-manager-5c8695d769-78g26\" (UID: \"572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b\") " pod="openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26" Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.348741 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-metrics-certs\") pod \"openstack-operator-controller-manager-5c8695d769-78g26\" (UID: \"572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b\") " pod="openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26" Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.348766 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98kvs\" (UniqueName: \"kubernetes.io/projected/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-kube-api-access-98kvs\") pod \"openstack-operator-controller-manager-5c8695d769-78g26\" (UID: \"572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b\") " pod="openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26" Dec 06 15:59:46 crc kubenswrapper[4813]: E1206 15:59:46.349344 4813 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 15:59:46 crc kubenswrapper[4813]: E1206 15:59:46.349383 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/279a2707-03fd-4cbc-9e13-d7a0d2a284db-cert podName:279a2707-03fd-4cbc-9e13-d7a0d2a284db nodeName:}" failed. No retries permitted until 2025-12-06 15:59:47.349370318 +0000 UTC m=+827.240249894 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/279a2707-03fd-4cbc-9e13-d7a0d2a284db-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fc7jrk" (UID: "279a2707-03fd-4cbc-9e13-d7a0d2a284db") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 15:59:46 crc kubenswrapper[4813]: E1206 15:59:46.349428 4813 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 15:59:46 crc kubenswrapper[4813]: E1206 15:59:46.349448 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-webhook-certs podName:572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b nodeName:}" failed. No retries permitted until 2025-12-06 15:59:46.849440849 +0000 UTC m=+826.740320425 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-webhook-certs") pod "openstack-operator-controller-manager-5c8695d769-78g26" (UID: "572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b") : secret "webhook-server-cert" not found Dec 06 15:59:46 crc kubenswrapper[4813]: E1206 15:59:46.349479 4813 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 15:59:46 crc kubenswrapper[4813]: E1206 15:59:46.349495 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-metrics-certs podName:572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b nodeName:}" failed. No retries permitted until 2025-12-06 15:59:46.849490451 +0000 UTC m=+826.740370027 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-metrics-certs") pod "openstack-operator-controller-manager-5c8695d769-78g26" (UID: "572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b") : secret "metrics-server-cert" not found Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.369898 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98kvs\" (UniqueName: \"kubernetes.io/projected/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-kube-api-access-98kvs\") pod \"openstack-operator-controller-manager-5c8695d769-78g26\" (UID: \"572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b\") " pod="openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26" Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.402782 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-g4ztj"] Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.452942 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqc5x\" (UniqueName: \"kubernetes.io/projected/00cfc20a-68d6-46be-98b8-83a43c9d11d9-kube-api-access-rqc5x\") pod \"rabbitmq-cluster-operator-manager-668c99d594-9lwhh\" (UID: \"00cfc20a-68d6-46be-98b8-83a43c9d11d9\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9lwhh" Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.459999 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-wxsqx"] Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.475705 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqc5x\" (UniqueName: \"kubernetes.io/projected/00cfc20a-68d6-46be-98b8-83a43c9d11d9-kube-api-access-rqc5x\") pod \"rabbitmq-cluster-operator-manager-668c99d594-9lwhh\" (UID: \"00cfc20a-68d6-46be-98b8-83a43c9d11d9\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9lwhh" Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.531901 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-f9nfd"] Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.554233 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6dc7dc95b4-6wznz"] Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.580537 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-8mt55"] Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.628442 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9lwhh" Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.732528 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-cjzht"] Dec 06 15:59:46 crc kubenswrapper[4813]: W1206 15:59:46.751468 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8870fd51_2141_4cc8_956b_050e849dcbd4.slice/crio-c955449d0f0a1fbee4a67429dd66eece5d02dba0afbef2677bc15ccc151ac340 WatchSource:0}: Error finding container c955449d0f0a1fbee4a67429dd66eece5d02dba0afbef2677bc15ccc151ac340: Status 404 returned error can't find the container with id c955449d0f0a1fbee4a67429dd66eece5d02dba0afbef2677bc15ccc151ac340 Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.755462 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-6rr2s"] Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.763528 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-6p6sf"] Dec 06 15:59:46 crc kubenswrapper[4813]: W1206 15:59:46.766450 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4cd1114_c317_4351_b566_05317dd589f1.slice/crio-8d0b4d286edb4707aa792faa5efa5efd405b9e9ba04ced47cbacd18cb3d44fbd WatchSource:0}: Error finding container 8d0b4d286edb4707aa792faa5efa5efd405b9e9ba04ced47cbacd18cb3d44fbd: Status 404 returned error can't find the container with id 8d0b4d286edb4707aa792faa5efa5efd405b9e9ba04ced47cbacd18cb3d44fbd Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.779727 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6983648f-60ef-429d-a2fd-de5dc7e6d0ba-cert\") pod \"infra-operator-controller-manager-78d48bff9d-pkvxj\" (UID: \"6983648f-60ef-429d-a2fd-de5dc7e6d0ba\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pkvxj" Dec 06 15:59:46 crc kubenswrapper[4813]: E1206 15:59:46.779904 4813 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 15:59:46 crc kubenswrapper[4813]: E1206 15:59:46.779946 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6983648f-60ef-429d-a2fd-de5dc7e6d0ba-cert podName:6983648f-60ef-429d-a2fd-de5dc7e6d0ba nodeName:}" failed. No retries permitted until 2025-12-06 15:59:48.779933007 +0000 UTC m=+828.670812583 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6983648f-60ef-429d-a2fd-de5dc7e6d0ba-cert") pod "infra-operator-controller-manager-78d48bff9d-pkvxj" (UID: "6983648f-60ef-429d-a2fd-de5dc7e6d0ba") : secret "infra-operator-webhook-server-cert" not found Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.880842 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-metrics-certs\") pod \"openstack-operator-controller-manager-5c8695d769-78g26\" (UID: \"572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b\") " pod="openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26" Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.880950 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-webhook-certs\") pod \"openstack-operator-controller-manager-5c8695d769-78g26\" (UID: \"572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b\") " pod="openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26" Dec 06 15:59:46 crc kubenswrapper[4813]: E1206 15:59:46.881053 4813 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 15:59:46 crc kubenswrapper[4813]: E1206 15:59:46.881102 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-webhook-certs podName:572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b nodeName:}" failed. No retries permitted until 2025-12-06 15:59:47.881089978 +0000 UTC m=+827.771969554 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-webhook-certs") pod "openstack-operator-controller-manager-5c8695d769-78g26" (UID: "572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b") : secret "webhook-server-cert" not found Dec 06 15:59:46 crc kubenswrapper[4813]: E1206 15:59:46.881142 4813 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 15:59:46 crc kubenswrapper[4813]: E1206 15:59:46.881161 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-metrics-certs podName:572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b nodeName:}" failed. No retries permitted until 2025-12-06 15:59:47.88115442 +0000 UTC m=+827.772033996 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-metrics-certs") pod "openstack-operator-controller-manager-5c8695d769-78g26" (UID: "572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b") : secret "metrics-server-cert" not found Dec 06 15:59:46 crc kubenswrapper[4813]: I1206 15:59:46.927453 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-cxh72"] Dec 06 15:59:46 crc kubenswrapper[4813]: W1206 15:59:46.934189 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d2f63b5_c93e_4949_bb79_459dad74dcf8.slice/crio-64791a3fb478e49a11a306cd01a7c16df76f2a6153349b9ce060d5886d4ad31c WatchSource:0}: Error finding container 64791a3fb478e49a11a306cd01a7c16df76f2a6153349b9ce060d5886d4ad31c: Status 404 returned error can't find the container with id 64791a3fb478e49a11a306cd01a7c16df76f2a6153349b9ce060d5886d4ad31c Dec 06 15:59:47 crc kubenswrapper[4813]: I1206 15:59:47.015581 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-g4ztj" event={"ID":"d67349e5-13fb-4859-ac8d-3e2c6bd67dbb","Type":"ContainerStarted","Data":"89ab0661919889de559cc756433f93a0a7c53a9d1c13793fedc4ebfe8e8b997f"} Dec 06 15:59:47 crc kubenswrapper[4813]: I1206 15:59:47.022759 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-4rdzx" event={"ID":"218be173-f41f-48b1-8af0-b94bbc483e8b","Type":"ContainerStarted","Data":"0b2b439c624d2af893b660101890e402ffa88120103e0f89de86b47076f0edf0"} Dec 06 15:59:47 crc kubenswrapper[4813]: I1206 15:59:47.023178 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-78fv6"] Dec 06 15:59:47 crc kubenswrapper[4813]: I1206 15:59:47.025344 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6dc7dc95b4-6wznz" event={"ID":"24d6dd0f-86e6-4b7a-beeb-56025a76ab5a","Type":"ContainerStarted","Data":"0cdaf16429713aee60a9802423a2b54d938f7277e14487a062c442cbd95cc60e"} Dec 06 15:59:47 crc kubenswrapper[4813]: I1206 15:59:47.026819 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-6p6sf" event={"ID":"a8763d41-9404-43ec-866b-9244bf2c4af9","Type":"ContainerStarted","Data":"cf1ad1d0e7d7acdaefeea6feb731684f191a9ef2d83208670d5d2471dc23b8a6"} Dec 06 15:59:47 crc kubenswrapper[4813]: I1206 15:59:47.037522 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-g9q96"] Dec 06 15:59:47 crc kubenswrapper[4813]: I1206 15:59:47.042591 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-2mpwn"] Dec 06 15:59:47 crc kubenswrapper[4813]: I1206 15:59:47.046557 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6n2hj"] Dec 06 15:59:47 crc kubenswrapper[4813]: W1206 15:59:47.056858 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb675acf0_51cc_43d9_b9c6_bd0ee1730d25.slice/crio-daea3fd45935afd183415330c80adbecb59b586dde3e3669bb160fd16a10268e WatchSource:0}: Error finding container daea3fd45935afd183415330c80adbecb59b586dde3e3669bb160fd16a10268e: Status 404 returned error can't find the container with id daea3fd45935afd183415330c80adbecb59b586dde3e3669bb160fd16a10268e Dec 06 15:59:47 crc kubenswrapper[4813]: I1206 15:59:47.069367 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-cjzht" event={"ID":"8870fd51-2141-4cc8-956b-050e849dcbd4","Type":"ContainerStarted","Data":"c955449d0f0a1fbee4a67429dd66eece5d02dba0afbef2677bc15ccc151ac340"} Dec 06 15:59:47 crc kubenswrapper[4813]: I1206 15:59:47.083432 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-cxh72" event={"ID":"4d2f63b5-c93e-4949-bb79-459dad74dcf8","Type":"ContainerStarted","Data":"64791a3fb478e49a11a306cd01a7c16df76f2a6153349b9ce060d5886d4ad31c"} Dec 06 15:59:47 crc kubenswrapper[4813]: I1206 15:59:47.085026 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-wxsqx" event={"ID":"e33ae276-0e3d-44b5-b970-17e98051c49c","Type":"ContainerStarted","Data":"e1c5356337de14801bd5f64173b47afd59bc8f256cc4a6937d8968a1595b2868"} Dec 06 15:59:47 crc kubenswrapper[4813]: I1206 15:59:47.085932 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8mt55" event={"ID":"a8b0c607-9997-44e8-a3cc-a36e4b2e70d2","Type":"ContainerStarted","Data":"1e5c39a07c77593d79155f26c498db22183d0ab98e3de9b754fbd2033e1a69a6"} Dec 06 15:59:47 crc kubenswrapper[4813]: I1206 15:59:47.087141 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-f9nfd" event={"ID":"3edd7ca8-a3ba-46c5-96c8-77aa9be672b7","Type":"ContainerStarted","Data":"5f1e71c75609babe897e68e3f352056976a955e2f6afff786867881065f85bf2"} Dec 06 15:59:47 crc kubenswrapper[4813]: I1206 15:59:47.092535 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-x8vkl"] Dec 06 15:59:47 crc kubenswrapper[4813]: I1206 15:59:47.106053 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-6rr2s" event={"ID":"b4cd1114-c317-4351-b566-05317dd589f1","Type":"ContainerStarted","Data":"8d0b4d286edb4707aa792faa5efa5efd405b9e9ba04ced47cbacd18cb3d44fbd"} Dec 06 15:59:47 crc kubenswrapper[4813]: I1206 15:59:47.112435 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qkrqw"] Dec 06 15:59:47 crc kubenswrapper[4813]: I1206 15:59:47.116030 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-rsbqf"] Dec 06 15:59:47 crc kubenswrapper[4813]: W1206 15:59:47.118388 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67c6b7ea_97cf_4ffa_9f3b_5111d4fbdeba.slice/crio-5452f0428e05025fa57e3fe08cc221147e822618eb0d8a6e633b1a05986869f3 WatchSource:0}: Error finding container 5452f0428e05025fa57e3fe08cc221147e822618eb0d8a6e633b1a05986869f3: Status 404 returned error can't find the container with id 5452f0428e05025fa57e3fe08cc221147e822618eb0d8a6e633b1a05986869f3 Dec 06 15:59:47 crc kubenswrapper[4813]: E1206 15:59:47.124636 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kmt5c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-58d5ff84df-qkrqw_openstack-operators(6d54c577-3fdf-48af-9070-301dd3b3eb81): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 15:59:47 crc kubenswrapper[4813]: E1206 15:59:47.126949 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kmt5c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-58d5ff84df-qkrqw_openstack-operators(6d54c577-3fdf-48af-9070-301dd3b3eb81): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 15:59:47 crc kubenswrapper[4813]: E1206 15:59:47.128592 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qkrqw" podUID="6d54c577-3fdf-48af-9070-301dd3b3eb81" Dec 06 15:59:47 crc kubenswrapper[4813]: E1206 15:59:47.129635 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-knrft,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-rsbqf_openstack-operators(67c6b7ea-97cf-4ffa-9f3b-5111d4fbdeba): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 15:59:47 crc kubenswrapper[4813]: E1206 15:59:47.129766 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bwj4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-9d58d64bc-x8vkl_openstack-operators(b05dafaa-0d81-4b80-ad3e-48da48a1655f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 15:59:47 crc kubenswrapper[4813]: E1206 15:59:47.137880 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bwj4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-9d58d64bc-x8vkl_openstack-operators(b05dafaa-0d81-4b80-ad3e-48da48a1655f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 15:59:47 crc kubenswrapper[4813]: E1206 15:59:47.138248 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-knrft,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-rsbqf_openstack-operators(67c6b7ea-97cf-4ffa-9f3b-5111d4fbdeba): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 15:59:47 crc kubenswrapper[4813]: E1206 15:59:47.139553 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-x8vkl" podUID="b05dafaa-0d81-4b80-ad3e-48da48a1655f" Dec 06 15:59:47 crc kubenswrapper[4813]: E1206 15:59:47.139661 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-rsbqf" podUID="67c6b7ea-97cf-4ffa-9f3b-5111d4fbdeba" Dec 06 15:59:47 crc kubenswrapper[4813]: I1206 15:59:47.184704 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-8zvf5"] Dec 06 15:59:47 crc kubenswrapper[4813]: I1206 15:59:47.192169 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-cbmwf"] Dec 06 15:59:47 crc kubenswrapper[4813]: E1206 15:59:47.194519 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8qswn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-667bd8d554-8zvf5_openstack-operators(5a8652b7-8a55-4658-acce-b7573421b94d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 15:59:47 crc kubenswrapper[4813]: E1206 15:59:47.196228 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8qswn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-667bd8d554-8zvf5_openstack-operators(5a8652b7-8a55-4658-acce-b7573421b94d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 15:59:47 crc kubenswrapper[4813]: E1206 15:59:47.197430 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-8zvf5" podUID="5a8652b7-8a55-4658-acce-b7573421b94d" Dec 06 15:59:47 crc kubenswrapper[4813]: W1206 15:59:47.214016 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod28acd95a_e657_46b5_a771_a24b27bdc6fe.slice/crio-65fb0495eb9b36d4527257107d875af190cf0f625e05abd8a54dc7b4c6148360 WatchSource:0}: Error finding container 65fb0495eb9b36d4527257107d875af190cf0f625e05abd8a54dc7b4c6148360: Status 404 returned error can't find the container with id 65fb0495eb9b36d4527257107d875af190cf0f625e05abd8a54dc7b4c6148360 Dec 06 15:59:47 crc kubenswrapper[4813]: E1206 15:59:47.216083 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9pn8x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-cbmwf_openstack-operators(28acd95a-e657-46b5-a771-a24b27bdc6fe): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 15:59:47 crc kubenswrapper[4813]: E1206 15:59:47.218549 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9pn8x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-cbmwf_openstack-operators(28acd95a-e657-46b5-a771-a24b27bdc6fe): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 15:59:47 crc kubenswrapper[4813]: E1206 15:59:47.220469 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-cbmwf" podUID="28acd95a-e657-46b5-a771-a24b27bdc6fe" Dec 06 15:59:47 crc kubenswrapper[4813]: I1206 15:59:47.286546 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9lwhh"] Dec 06 15:59:47 crc kubenswrapper[4813]: W1206 15:59:47.287967 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00cfc20a_68d6_46be_98b8_83a43c9d11d9.slice/crio-413cf2e91ba9083542ae49818b19274c69e33fd6b47d9e9fda5657380ae54f10 WatchSource:0}: Error finding container 413cf2e91ba9083542ae49818b19274c69e33fd6b47d9e9fda5657380ae54f10: Status 404 returned error can't find the container with id 413cf2e91ba9083542ae49818b19274c69e33fd6b47d9e9fda5657380ae54f10 Dec 06 15:59:47 crc kubenswrapper[4813]: I1206 15:59:47.386868 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/279a2707-03fd-4cbc-9e13-d7a0d2a284db-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fc7jrk\" (UID: \"279a2707-03fd-4cbc-9e13-d7a0d2a284db\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fc7jrk" Dec 06 15:59:47 crc kubenswrapper[4813]: E1206 15:59:47.387014 4813 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 15:59:47 crc kubenswrapper[4813]: E1206 15:59:47.387070 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/279a2707-03fd-4cbc-9e13-d7a0d2a284db-cert podName:279a2707-03fd-4cbc-9e13-d7a0d2a284db nodeName:}" failed. No retries permitted until 2025-12-06 15:59:49.387054918 +0000 UTC m=+829.277934484 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/279a2707-03fd-4cbc-9e13-d7a0d2a284db-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fc7jrk" (UID: "279a2707-03fd-4cbc-9e13-d7a0d2a284db") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 15:59:47 crc kubenswrapper[4813]: I1206 15:59:47.897154 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-metrics-certs\") pod \"openstack-operator-controller-manager-5c8695d769-78g26\" (UID: \"572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b\") " pod="openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26" Dec 06 15:59:47 crc kubenswrapper[4813]: I1206 15:59:47.897276 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-webhook-certs\") pod \"openstack-operator-controller-manager-5c8695d769-78g26\" (UID: \"572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b\") " pod="openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26" Dec 06 15:59:47 crc kubenswrapper[4813]: E1206 15:59:47.897295 4813 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 15:59:47 crc kubenswrapper[4813]: E1206 15:59:47.897371 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-metrics-certs podName:572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b nodeName:}" failed. No retries permitted until 2025-12-06 15:59:49.897341152 +0000 UTC m=+829.788220728 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-metrics-certs") pod "openstack-operator-controller-manager-5c8695d769-78g26" (UID: "572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b") : secret "metrics-server-cert" not found Dec 06 15:59:47 crc kubenswrapper[4813]: E1206 15:59:47.897416 4813 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 15:59:47 crc kubenswrapper[4813]: E1206 15:59:47.897465 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-webhook-certs podName:572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b nodeName:}" failed. No retries permitted until 2025-12-06 15:59:49.897450615 +0000 UTC m=+829.788330181 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-webhook-certs") pod "openstack-operator-controller-manager-5c8695d769-78g26" (UID: "572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b") : secret "webhook-server-cert" not found Dec 06 15:59:48 crc kubenswrapper[4813]: I1206 15:59:48.128871 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-g9q96" event={"ID":"031a9bd1-298a-4945-a52d-452005c8467e","Type":"ContainerStarted","Data":"58ae63a2e0a3c3e8c0658f57ef7f275c636d3095e8e3156a8ccca5830c2ed444"} Dec 06 15:59:48 crc kubenswrapper[4813]: I1206 15:59:48.131686 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-8zvf5" event={"ID":"5a8652b7-8a55-4658-acce-b7573421b94d","Type":"ContainerStarted","Data":"9d22355ffc7d01d5ac7e66eb932aae14b02bf857a4969f00d4379e77177742e9"} Dec 06 15:59:48 crc kubenswrapper[4813]: E1206 15:59:48.136593 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-8zvf5" podUID="5a8652b7-8a55-4658-acce-b7573421b94d" Dec 06 15:59:48 crc kubenswrapper[4813]: I1206 15:59:48.138366 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-cbmwf" event={"ID":"28acd95a-e657-46b5-a771-a24b27bdc6fe","Type":"ContainerStarted","Data":"65fb0495eb9b36d4527257107d875af190cf0f625e05abd8a54dc7b4c6148360"} Dec 06 15:59:48 crc kubenswrapper[4813]: I1206 15:59:48.146197 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-x8vkl" event={"ID":"b05dafaa-0d81-4b80-ad3e-48da48a1655f","Type":"ContainerStarted","Data":"3063874783cd9c03c21811f1aa9fc5781f7e50198f75c729bd3752c3ddad9196"} Dec 06 15:59:48 crc kubenswrapper[4813]: I1206 15:59:48.150417 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6n2hj" event={"ID":"99a0eae7-c990-42cd-a18e-78b81177bbb6","Type":"ContainerStarted","Data":"1bf9cc4e976140846e205696b5d8a5d3dcbfb34fd5f0f9b90229fdb9487cf711"} Dec 06 15:59:48 crc kubenswrapper[4813]: E1206 15:59:48.158078 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-cbmwf" podUID="28acd95a-e657-46b5-a771-a24b27bdc6fe" Dec 06 15:59:48 crc kubenswrapper[4813]: I1206 15:59:48.159482 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qkrqw" event={"ID":"6d54c577-3fdf-48af-9070-301dd3b3eb81","Type":"ContainerStarted","Data":"5304820ac674c6ec3a3d82ca29e880946306f0946adb78e0313b2499a5845cdf"} Dec 06 15:59:48 crc kubenswrapper[4813]: E1206 15:59:48.159792 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-x8vkl" podUID="b05dafaa-0d81-4b80-ad3e-48da48a1655f" Dec 06 15:59:48 crc kubenswrapper[4813]: I1206 15:59:48.165786 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-2mpwn" event={"ID":"c35efae6-9bf9-466d-ac8d-203d8a2c4b97","Type":"ContainerStarted","Data":"6475deff7700b327fc1dc66ec77ab63b21c45fa0de1ce2ca3f70a89c8e3fa87c"} Dec 06 15:59:48 crc kubenswrapper[4813]: I1206 15:59:48.170008 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9lwhh" event={"ID":"00cfc20a-68d6-46be-98b8-83a43c9d11d9","Type":"ContainerStarted","Data":"413cf2e91ba9083542ae49818b19274c69e33fd6b47d9e9fda5657380ae54f10"} Dec 06 15:59:48 crc kubenswrapper[4813]: I1206 15:59:48.172783 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-78fv6" event={"ID":"b675acf0-51cc-43d9-b9c6-bd0ee1730d25","Type":"ContainerStarted","Data":"daea3fd45935afd183415330c80adbecb59b586dde3e3669bb160fd16a10268e"} Dec 06 15:59:48 crc kubenswrapper[4813]: E1206 15:59:48.173209 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qkrqw" podUID="6d54c577-3fdf-48af-9070-301dd3b3eb81" Dec 06 15:59:48 crc kubenswrapper[4813]: I1206 15:59:48.196538 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-rsbqf" event={"ID":"67c6b7ea-97cf-4ffa-9f3b-5111d4fbdeba","Type":"ContainerStarted","Data":"5452f0428e05025fa57e3fe08cc221147e822618eb0d8a6e633b1a05986869f3"} Dec 06 15:59:48 crc kubenswrapper[4813]: E1206 15:59:48.203172 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-rsbqf" podUID="67c6b7ea-97cf-4ffa-9f3b-5111d4fbdeba" Dec 06 15:59:48 crc kubenswrapper[4813]: I1206 15:59:48.811963 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6983648f-60ef-429d-a2fd-de5dc7e6d0ba-cert\") pod \"infra-operator-controller-manager-78d48bff9d-pkvxj\" (UID: \"6983648f-60ef-429d-a2fd-de5dc7e6d0ba\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pkvxj" Dec 06 15:59:48 crc kubenswrapper[4813]: E1206 15:59:48.812162 4813 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 15:59:48 crc kubenswrapper[4813]: E1206 15:59:48.812244 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6983648f-60ef-429d-a2fd-de5dc7e6d0ba-cert podName:6983648f-60ef-429d-a2fd-de5dc7e6d0ba nodeName:}" failed. No retries permitted until 2025-12-06 15:59:52.812225701 +0000 UTC m=+832.703105277 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6983648f-60ef-429d-a2fd-de5dc7e6d0ba-cert") pod "infra-operator-controller-manager-78d48bff9d-pkvxj" (UID: "6983648f-60ef-429d-a2fd-de5dc7e6d0ba") : secret "infra-operator-webhook-server-cert" not found Dec 06 15:59:49 crc kubenswrapper[4813]: E1206 15:59:49.213778 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-cbmwf" podUID="28acd95a-e657-46b5-a771-a24b27bdc6fe" Dec 06 15:59:49 crc kubenswrapper[4813]: E1206 15:59:49.215103 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-x8vkl" podUID="b05dafaa-0d81-4b80-ad3e-48da48a1655f" Dec 06 15:59:49 crc kubenswrapper[4813]: E1206 15:59:49.215250 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qkrqw" podUID="6d54c577-3fdf-48af-9070-301dd3b3eb81" Dec 06 15:59:49 crc kubenswrapper[4813]: E1206 15:59:49.215844 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-8zvf5" podUID="5a8652b7-8a55-4658-acce-b7573421b94d" Dec 06 15:59:49 crc kubenswrapper[4813]: E1206 15:59:49.215888 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-rsbqf" podUID="67c6b7ea-97cf-4ffa-9f3b-5111d4fbdeba" Dec 06 15:59:49 crc kubenswrapper[4813]: I1206 15:59:49.421433 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/279a2707-03fd-4cbc-9e13-d7a0d2a284db-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fc7jrk\" (UID: \"279a2707-03fd-4cbc-9e13-d7a0d2a284db\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fc7jrk" Dec 06 15:59:49 crc kubenswrapper[4813]: E1206 15:59:49.421592 4813 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 15:59:49 crc kubenswrapper[4813]: E1206 15:59:49.421651 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/279a2707-03fd-4cbc-9e13-d7a0d2a284db-cert podName:279a2707-03fd-4cbc-9e13-d7a0d2a284db nodeName:}" failed. No retries permitted until 2025-12-06 15:59:53.421637153 +0000 UTC m=+833.312516729 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/279a2707-03fd-4cbc-9e13-d7a0d2a284db-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fc7jrk" (UID: "279a2707-03fd-4cbc-9e13-d7a0d2a284db") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 15:59:49 crc kubenswrapper[4813]: I1206 15:59:49.427436 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 15:59:49 crc kubenswrapper[4813]: I1206 15:59:49.427472 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 15:59:49 crc kubenswrapper[4813]: I1206 15:59:49.427504 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" Dec 06 15:59:49 crc kubenswrapper[4813]: I1206 15:59:49.428008 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0d0cad4af55e6ecbb2a20d8a543c0848e24b8077a79dec0b4034636d4a17f514"} pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 15:59:49 crc kubenswrapper[4813]: I1206 15:59:49.428060 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" containerID="cri-o://0d0cad4af55e6ecbb2a20d8a543c0848e24b8077a79dec0b4034636d4a17f514" gracePeriod=600 Dec 06 15:59:49 crc kubenswrapper[4813]: I1206 15:59:49.928281 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-metrics-certs\") pod \"openstack-operator-controller-manager-5c8695d769-78g26\" (UID: \"572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b\") " pod="openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26" Dec 06 15:59:49 crc kubenswrapper[4813]: I1206 15:59:49.928578 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-webhook-certs\") pod \"openstack-operator-controller-manager-5c8695d769-78g26\" (UID: \"572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b\") " pod="openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26" Dec 06 15:59:49 crc kubenswrapper[4813]: E1206 15:59:49.928716 4813 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 15:59:49 crc kubenswrapper[4813]: E1206 15:59:49.928759 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-webhook-certs podName:572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b nodeName:}" failed. No retries permitted until 2025-12-06 15:59:53.928746014 +0000 UTC m=+833.819625580 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-webhook-certs") pod "openstack-operator-controller-manager-5c8695d769-78g26" (UID: "572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b") : secret "webhook-server-cert" not found Dec 06 15:59:49 crc kubenswrapper[4813]: E1206 15:59:49.929069 4813 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 15:59:49 crc kubenswrapper[4813]: E1206 15:59:49.929097 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-metrics-certs podName:572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b nodeName:}" failed. No retries permitted until 2025-12-06 15:59:53.929090443 +0000 UTC m=+833.819970019 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-metrics-certs") pod "openstack-operator-controller-manager-5c8695d769-78g26" (UID: "572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b") : secret "metrics-server-cert" not found Dec 06 15:59:50 crc kubenswrapper[4813]: I1206 15:59:50.235076 4813 generic.go:334] "Generic (PLEG): container finished" podID="d88e8bae-c055-4c55-b548-f621ff96de06" containerID="0d0cad4af55e6ecbb2a20d8a543c0848e24b8077a79dec0b4034636d4a17f514" exitCode=0 Dec 06 15:59:50 crc kubenswrapper[4813]: I1206 15:59:50.235115 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerDied","Data":"0d0cad4af55e6ecbb2a20d8a543c0848e24b8077a79dec0b4034636d4a17f514"} Dec 06 15:59:50 crc kubenswrapper[4813]: I1206 15:59:50.235149 4813 scope.go:117] "RemoveContainer" containerID="6841d4ac1913266f58304ddc35c3d4b57d8eed36652e13d503003868ba785be0" Dec 06 15:59:52 crc kubenswrapper[4813]: I1206 15:59:52.868846 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6983648f-60ef-429d-a2fd-de5dc7e6d0ba-cert\") pod \"infra-operator-controller-manager-78d48bff9d-pkvxj\" (UID: \"6983648f-60ef-429d-a2fd-de5dc7e6d0ba\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pkvxj" Dec 06 15:59:52 crc kubenswrapper[4813]: E1206 15:59:52.869023 4813 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 15:59:52 crc kubenswrapper[4813]: E1206 15:59:52.869212 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6983648f-60ef-429d-a2fd-de5dc7e6d0ba-cert podName:6983648f-60ef-429d-a2fd-de5dc7e6d0ba nodeName:}" failed. No retries permitted until 2025-12-06 16:00:00.869181538 +0000 UTC m=+840.760061114 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6983648f-60ef-429d-a2fd-de5dc7e6d0ba-cert") pod "infra-operator-controller-manager-78d48bff9d-pkvxj" (UID: "6983648f-60ef-429d-a2fd-de5dc7e6d0ba") : secret "infra-operator-webhook-server-cert" not found Dec 06 15:59:53 crc kubenswrapper[4813]: I1206 15:59:53.478445 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/279a2707-03fd-4cbc-9e13-d7a0d2a284db-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fc7jrk\" (UID: \"279a2707-03fd-4cbc-9e13-d7a0d2a284db\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fc7jrk" Dec 06 15:59:53 crc kubenswrapper[4813]: E1206 15:59:53.478653 4813 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 15:59:53 crc kubenswrapper[4813]: E1206 15:59:53.478744 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/279a2707-03fd-4cbc-9e13-d7a0d2a284db-cert podName:279a2707-03fd-4cbc-9e13-d7a0d2a284db nodeName:}" failed. No retries permitted until 2025-12-06 16:00:01.478721533 +0000 UTC m=+841.369601119 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/279a2707-03fd-4cbc-9e13-d7a0d2a284db-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fc7jrk" (UID: "279a2707-03fd-4cbc-9e13-d7a0d2a284db") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 15:59:53 crc kubenswrapper[4813]: I1206 15:59:53.985327 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-webhook-certs\") pod \"openstack-operator-controller-manager-5c8695d769-78g26\" (UID: \"572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b\") " pod="openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26" Dec 06 15:59:53 crc kubenswrapper[4813]: I1206 15:59:53.985426 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-metrics-certs\") pod \"openstack-operator-controller-manager-5c8695d769-78g26\" (UID: \"572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b\") " pod="openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26" Dec 06 15:59:53 crc kubenswrapper[4813]: E1206 15:59:53.985565 4813 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 15:59:53 crc kubenswrapper[4813]: E1206 15:59:53.985593 4813 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 15:59:53 crc kubenswrapper[4813]: E1206 15:59:53.985631 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-metrics-certs podName:572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b nodeName:}" failed. No retries permitted until 2025-12-06 16:00:01.985612729 +0000 UTC m=+841.876492315 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-metrics-certs") pod "openstack-operator-controller-manager-5c8695d769-78g26" (UID: "572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b") : secret "metrics-server-cert" not found Dec 06 15:59:53 crc kubenswrapper[4813]: E1206 15:59:53.985685 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-webhook-certs podName:572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b nodeName:}" failed. No retries permitted until 2025-12-06 16:00:01.98566107 +0000 UTC m=+841.876540746 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-webhook-certs") pod "openstack-operator-controller-manager-5c8695d769-78g26" (UID: "572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b") : secret "webhook-server-cert" not found Dec 06 16:00:00 crc kubenswrapper[4813]: I1206 16:00:00.158424 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417280-bd6bw"] Dec 06 16:00:00 crc kubenswrapper[4813]: I1206 16:00:00.159555 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417280-bd6bw" Dec 06 16:00:00 crc kubenswrapper[4813]: I1206 16:00:00.163463 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 16:00:00 crc kubenswrapper[4813]: I1206 16:00:00.163618 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 16:00:00 crc kubenswrapper[4813]: I1206 16:00:00.171228 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417280-bd6bw"] Dec 06 16:00:00 crc kubenswrapper[4813]: I1206 16:00:00.184322 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/971a4ad6-5f62-47bc-b836-b7e6faaae668-secret-volume\") pod \"collect-profiles-29417280-bd6bw\" (UID: \"971a4ad6-5f62-47bc-b836-b7e6faaae668\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417280-bd6bw" Dec 06 16:00:00 crc kubenswrapper[4813]: I1206 16:00:00.184402 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fjdk\" (UniqueName: \"kubernetes.io/projected/971a4ad6-5f62-47bc-b836-b7e6faaae668-kube-api-access-8fjdk\") pod \"collect-profiles-29417280-bd6bw\" (UID: \"971a4ad6-5f62-47bc-b836-b7e6faaae668\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417280-bd6bw" Dec 06 16:00:00 crc kubenswrapper[4813]: I1206 16:00:00.184462 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/971a4ad6-5f62-47bc-b836-b7e6faaae668-config-volume\") pod \"collect-profiles-29417280-bd6bw\" (UID: \"971a4ad6-5f62-47bc-b836-b7e6faaae668\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417280-bd6bw" Dec 06 16:00:00 crc kubenswrapper[4813]: E1206 16:00:00.225365 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.74:5001/openstack-k8s-operators/barbican-operator:050849d6558b19b4244a31c408c5335e92d1a054" Dec 06 16:00:00 crc kubenswrapper[4813]: E1206 16:00:00.225414 4813 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.74:5001/openstack-k8s-operators/barbican-operator:050849d6558b19b4244a31c408c5335e92d1a054" Dec 06 16:00:00 crc kubenswrapper[4813]: E1206 16:00:00.225541 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.74:5001/openstack-k8s-operators/barbican-operator:050849d6558b19b4244a31c408c5335e92d1a054,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fjq65,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-6dc7dc95b4-6wznz_openstack-operators(24d6dd0f-86e6-4b7a-beeb-56025a76ab5a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 16:00:00 crc kubenswrapper[4813]: I1206 16:00:00.287080 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/971a4ad6-5f62-47bc-b836-b7e6faaae668-secret-volume\") pod \"collect-profiles-29417280-bd6bw\" (UID: \"971a4ad6-5f62-47bc-b836-b7e6faaae668\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417280-bd6bw" Dec 06 16:00:00 crc kubenswrapper[4813]: I1206 16:00:00.287153 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fjdk\" (UniqueName: \"kubernetes.io/projected/971a4ad6-5f62-47bc-b836-b7e6faaae668-kube-api-access-8fjdk\") pod \"collect-profiles-29417280-bd6bw\" (UID: \"971a4ad6-5f62-47bc-b836-b7e6faaae668\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417280-bd6bw" Dec 06 16:00:00 crc kubenswrapper[4813]: I1206 16:00:00.287197 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/971a4ad6-5f62-47bc-b836-b7e6faaae668-config-volume\") pod \"collect-profiles-29417280-bd6bw\" (UID: \"971a4ad6-5f62-47bc-b836-b7e6faaae668\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417280-bd6bw" Dec 06 16:00:00 crc kubenswrapper[4813]: I1206 16:00:00.288982 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/971a4ad6-5f62-47bc-b836-b7e6faaae668-config-volume\") pod \"collect-profiles-29417280-bd6bw\" (UID: \"971a4ad6-5f62-47bc-b836-b7e6faaae668\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417280-bd6bw" Dec 06 16:00:00 crc kubenswrapper[4813]: I1206 16:00:00.294356 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/971a4ad6-5f62-47bc-b836-b7e6faaae668-secret-volume\") pod \"collect-profiles-29417280-bd6bw\" (UID: \"971a4ad6-5f62-47bc-b836-b7e6faaae668\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417280-bd6bw" Dec 06 16:00:00 crc kubenswrapper[4813]: I1206 16:00:00.307639 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fjdk\" (UniqueName: \"kubernetes.io/projected/971a4ad6-5f62-47bc-b836-b7e6faaae668-kube-api-access-8fjdk\") pod \"collect-profiles-29417280-bd6bw\" (UID: \"971a4ad6-5f62-47bc-b836-b7e6faaae668\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417280-bd6bw" Dec 06 16:00:00 crc kubenswrapper[4813]: I1206 16:00:00.485610 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 16:00:00 crc kubenswrapper[4813]: I1206 16:00:00.493815 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417280-bd6bw" Dec 06 16:00:00 crc kubenswrapper[4813]: I1206 16:00:00.894580 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6983648f-60ef-429d-a2fd-de5dc7e6d0ba-cert\") pod \"infra-operator-controller-manager-78d48bff9d-pkvxj\" (UID: \"6983648f-60ef-429d-a2fd-de5dc7e6d0ba\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pkvxj" Dec 06 16:00:00 crc kubenswrapper[4813]: E1206 16:00:00.894718 4813 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 16:00:00 crc kubenswrapper[4813]: E1206 16:00:00.895000 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6983648f-60ef-429d-a2fd-de5dc7e6d0ba-cert podName:6983648f-60ef-429d-a2fd-de5dc7e6d0ba nodeName:}" failed. No retries permitted until 2025-12-06 16:00:16.894985385 +0000 UTC m=+856.785864961 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6983648f-60ef-429d-a2fd-de5dc7e6d0ba-cert") pod "infra-operator-controller-manager-78d48bff9d-pkvxj" (UID: "6983648f-60ef-429d-a2fd-de5dc7e6d0ba") : secret "infra-operator-webhook-server-cert" not found Dec 06 16:00:01 crc kubenswrapper[4813]: E1206 16:00:01.048230 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5" Dec 06 16:00:01 crc kubenswrapper[4813]: E1206 16:00:01.048415 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zptcz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-g4ztj_openstack-operators(d67349e5-13fb-4859-ac8d-3e2c6bd67dbb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 16:00:01 crc kubenswrapper[4813]: I1206 16:00:01.513361 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/279a2707-03fd-4cbc-9e13-d7a0d2a284db-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fc7jrk\" (UID: \"279a2707-03fd-4cbc-9e13-d7a0d2a284db\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fc7jrk" Dec 06 16:00:01 crc kubenswrapper[4813]: E1206 16:00:01.514453 4813 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 16:00:01 crc kubenswrapper[4813]: E1206 16:00:01.514553 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/279a2707-03fd-4cbc-9e13-d7a0d2a284db-cert podName:279a2707-03fd-4cbc-9e13-d7a0d2a284db nodeName:}" failed. No retries permitted until 2025-12-06 16:00:17.514521513 +0000 UTC m=+857.405401089 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/279a2707-03fd-4cbc-9e13-d7a0d2a284db-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fc7jrk" (UID: "279a2707-03fd-4cbc-9e13-d7a0d2a284db") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 16:00:02 crc kubenswrapper[4813]: I1206 16:00:02.018538 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-metrics-certs\") pod \"openstack-operator-controller-manager-5c8695d769-78g26\" (UID: \"572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b\") " pod="openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26" Dec 06 16:00:02 crc kubenswrapper[4813]: I1206 16:00:02.018637 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-webhook-certs\") pod \"openstack-operator-controller-manager-5c8695d769-78g26\" (UID: \"572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b\") " pod="openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26" Dec 06 16:00:02 crc kubenswrapper[4813]: E1206 16:00:02.018747 4813 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 16:00:02 crc kubenswrapper[4813]: E1206 16:00:02.018827 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-metrics-certs podName:572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b nodeName:}" failed. No retries permitted until 2025-12-06 16:00:18.01881004 +0000 UTC m=+857.909689616 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-metrics-certs") pod "openstack-operator-controller-manager-5c8695d769-78g26" (UID: "572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b") : secret "metrics-server-cert" not found Dec 06 16:00:02 crc kubenswrapper[4813]: E1206 16:00:02.018759 4813 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 16:00:02 crc kubenswrapper[4813]: E1206 16:00:02.019160 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-webhook-certs podName:572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b nodeName:}" failed. No retries permitted until 2025-12-06 16:00:18.019144689 +0000 UTC m=+857.910024265 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-webhook-certs") pod "openstack-operator-controller-manager-5c8695d769-78g26" (UID: "572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b") : secret "webhook-server-cert" not found Dec 06 16:00:02 crc kubenswrapper[4813]: E1206 16:00:02.316553 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59" Dec 06 16:00:02 crc kubenswrapper[4813]: E1206 16:00:02.316938 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xbdgh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-2mpwn_openstack-operators(c35efae6-9bf9-466d-ac8d-203d8a2c4b97): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 16:00:04 crc kubenswrapper[4813]: E1206 16:00:04.362347 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:981b6a8f95934a86c5f10ef6e198b07265aeba7f11cf84b9ccd13dfaf06f3ca3" Dec 06 16:00:04 crc kubenswrapper[4813]: E1206 16:00:04.362651 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:981b6a8f95934a86c5f10ef6e198b07265aeba7f11cf84b9ccd13dfaf06f3ca3,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-h9mpn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-6c677c69b-8mt55_openstack-operators(a8b0c607-9997-44e8-a3cc-a36e4b2e70d2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 16:00:07 crc kubenswrapper[4813]: E1206 16:00:07.182140 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429" Dec 06 16:00:07 crc kubenswrapper[4813]: E1206 16:00:07.183936 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7p57p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-wxsqx_openstack-operators(e33ae276-0e3d-44b5-b970-17e98051c49c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 16:00:11 crc kubenswrapper[4813]: E1206 16:00:11.188997 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad" Dec 06 16:00:11 crc kubenswrapper[4813]: E1206 16:00:11.189466 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7z7fw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-79c8c4686c-6n2hj_openstack-operators(99a0eae7-c990-42cd-a18e-78b81177bbb6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 16:00:11 crc kubenswrapper[4813]: E1206 16:00:11.765036 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 06 16:00:11 crc kubenswrapper[4813]: E1206 16:00:11.765204 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-knwhm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-cjzht_openstack-operators(8870fd51-2141-4cc8-956b-050e849dcbd4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 16:00:16 crc kubenswrapper[4813]: E1206 16:00:16.306413 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 06 16:00:16 crc kubenswrapper[4813]: E1206 16:00:16.307217 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hcdhd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-78fv6_openstack-operators(b675acf0-51cc-43d9-b9c6-bd0ee1730d25): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 16:00:16 crc kubenswrapper[4813]: E1206 16:00:16.750506 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 06 16:00:16 crc kubenswrapper[4813]: E1206 16:00:16.750672 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rqc5x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-9lwhh_openstack-operators(00cfc20a-68d6-46be-98b8-83a43c9d11d9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 16:00:16 crc kubenswrapper[4813]: E1206 16:00:16.751999 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9lwhh" podUID="00cfc20a-68d6-46be-98b8-83a43c9d11d9" Dec 06 16:00:16 crc kubenswrapper[4813]: I1206 16:00:16.954233 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6983648f-60ef-429d-a2fd-de5dc7e6d0ba-cert\") pod \"infra-operator-controller-manager-78d48bff9d-pkvxj\" (UID: \"6983648f-60ef-429d-a2fd-de5dc7e6d0ba\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pkvxj" Dec 06 16:00:16 crc kubenswrapper[4813]: I1206 16:00:16.973082 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6983648f-60ef-429d-a2fd-de5dc7e6d0ba-cert\") pod \"infra-operator-controller-manager-78d48bff9d-pkvxj\" (UID: \"6983648f-60ef-429d-a2fd-de5dc7e6d0ba\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pkvxj" Dec 06 16:00:17 crc kubenswrapper[4813]: I1206 16:00:17.193890 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-l9rd4" Dec 06 16:00:17 crc kubenswrapper[4813]: I1206 16:00:17.202763 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pkvxj" Dec 06 16:00:17 crc kubenswrapper[4813]: E1206 16:00:17.503700 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9lwhh" podUID="00cfc20a-68d6-46be-98b8-83a43c9d11d9" Dec 06 16:00:17 crc kubenswrapper[4813]: I1206 16:00:17.561563 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/279a2707-03fd-4cbc-9e13-d7a0d2a284db-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fc7jrk\" (UID: \"279a2707-03fd-4cbc-9e13-d7a0d2a284db\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fc7jrk" Dec 06 16:00:17 crc kubenswrapper[4813]: I1206 16:00:17.579159 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/279a2707-03fd-4cbc-9e13-d7a0d2a284db-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fc7jrk\" (UID: \"279a2707-03fd-4cbc-9e13-d7a0d2a284db\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fc7jrk" Dec 06 16:00:17 crc kubenswrapper[4813]: I1206 16:00:17.660768 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-rrd7f" Dec 06 16:00:17 crc kubenswrapper[4813]: I1206 16:00:17.669083 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fc7jrk" Dec 06 16:00:18 crc kubenswrapper[4813]: I1206 16:00:18.066040 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-webhook-certs\") pod \"openstack-operator-controller-manager-5c8695d769-78g26\" (UID: \"572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b\") " pod="openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26" Dec 06 16:00:18 crc kubenswrapper[4813]: I1206 16:00:18.066355 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-metrics-certs\") pod \"openstack-operator-controller-manager-5c8695d769-78g26\" (UID: \"572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b\") " pod="openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26" Dec 06 16:00:18 crc kubenswrapper[4813]: I1206 16:00:18.072387 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-metrics-certs\") pod \"openstack-operator-controller-manager-5c8695d769-78g26\" (UID: \"572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b\") " pod="openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26" Dec 06 16:00:18 crc kubenswrapper[4813]: I1206 16:00:18.072426 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b-webhook-certs\") pod \"openstack-operator-controller-manager-5c8695d769-78g26\" (UID: \"572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b\") " pod="openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26" Dec 06 16:00:18 crc kubenswrapper[4813]: I1206 16:00:18.099897 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-t6rxr" Dec 06 16:00:18 crc kubenswrapper[4813]: I1206 16:00:18.108648 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26" Dec 06 16:00:18 crc kubenswrapper[4813]: I1206 16:00:18.512577 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerStarted","Data":"659cf1c13b1c009e3dc025bbc1b70aff2561005d40bd452ea21f924d8a1cca89"} Dec 06 16:00:18 crc kubenswrapper[4813]: I1206 16:00:18.553538 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417280-bd6bw"] Dec 06 16:00:19 crc kubenswrapper[4813]: W1206 16:00:19.645982 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod971a4ad6_5f62_47bc_b836_b7e6faaae668.slice/crio-60aa677834ba1c5954e002df4b024349e8a0c50a3b67e634a5102ff1235e5fa3 WatchSource:0}: Error finding container 60aa677834ba1c5954e002df4b024349e8a0c50a3b67e634a5102ff1235e5fa3: Status 404 returned error can't find the container with id 60aa677834ba1c5954e002df4b024349e8a0c50a3b67e634a5102ff1235e5fa3 Dec 06 16:00:20 crc kubenswrapper[4813]: I1206 16:00:20.044608 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fc7jrk"] Dec 06 16:00:20 crc kubenswrapper[4813]: I1206 16:00:20.240720 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-pkvxj"] Dec 06 16:00:20 crc kubenswrapper[4813]: I1206 16:00:20.335965 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26"] Dec 06 16:00:20 crc kubenswrapper[4813]: I1206 16:00:20.535125 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-4rdzx" event={"ID":"218be173-f41f-48b1-8af0-b94bbc483e8b","Type":"ContainerStarted","Data":"00a54341e849b52d5559fb115e2f9f9da279805d8abe159a3db5c28e640c0342"} Dec 06 16:00:20 crc kubenswrapper[4813]: I1206 16:00:20.540123 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-f9nfd" event={"ID":"3edd7ca8-a3ba-46c5-96c8-77aa9be672b7","Type":"ContainerStarted","Data":"b5fcdf87abd0e7fc0991ca0f5a8ee4ed9fd32a2490c8e6b41fb45a35a3a943d7"} Dec 06 16:00:20 crc kubenswrapper[4813]: I1206 16:00:20.542870 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-6p6sf" event={"ID":"a8763d41-9404-43ec-866b-9244bf2c4af9","Type":"ContainerStarted","Data":"a678b467c87c86848e93b51d1f81c446ab83fb99efa78036d65eacb059090218"} Dec 06 16:00:20 crc kubenswrapper[4813]: I1206 16:00:20.543790 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417280-bd6bw" event={"ID":"971a4ad6-5f62-47bc-b836-b7e6faaae668","Type":"ContainerStarted","Data":"60aa677834ba1c5954e002df4b024349e8a0c50a3b67e634a5102ff1235e5fa3"} Dec 06 16:00:20 crc kubenswrapper[4813]: I1206 16:00:20.544912 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-cxh72" event={"ID":"4d2f63b5-c93e-4949-bb79-459dad74dcf8","Type":"ContainerStarted","Data":"05472ba22c40fc3e8b03151d2156093905a69a0adbacd63642c9e54f9aa2a189"} Dec 06 16:00:21 crc kubenswrapper[4813]: I1206 16:00:21.558646 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pkvxj" event={"ID":"6983648f-60ef-429d-a2fd-de5dc7e6d0ba","Type":"ContainerStarted","Data":"58504f793af168de5eb7b848e41a8f741eeb84a1d265b8ffba2ebe5a5deb724b"} Dec 06 16:00:21 crc kubenswrapper[4813]: I1206 16:00:21.561790 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-6rr2s" event={"ID":"b4cd1114-c317-4351-b566-05317dd589f1","Type":"ContainerStarted","Data":"0934df0553b17b8fb10bf21a2ceb983e098ce347309cde586c6a6f95f102b9ee"} Dec 06 16:00:21 crc kubenswrapper[4813]: I1206 16:00:21.563668 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-g9q96" event={"ID":"031a9bd1-298a-4945-a52d-452005c8467e","Type":"ContainerStarted","Data":"68f577250abbea4eeae49ca111f2ef046cd54c03cbb2fbb31cb885cb4fd58232"} Dec 06 16:00:21 crc kubenswrapper[4813]: I1206 16:00:21.564510 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fc7jrk" event={"ID":"279a2707-03fd-4cbc-9e13-d7a0d2a284db","Type":"ContainerStarted","Data":"d56b6093d10314420abe6b31b45496c12d1de3e513e019c5ed41fc1ce127dea7"} Dec 06 16:00:21 crc kubenswrapper[4813]: I1206 16:00:21.566455 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-x8vkl" event={"ID":"b05dafaa-0d81-4b80-ad3e-48da48a1655f","Type":"ContainerStarted","Data":"9654b8ba4b1132e22eccf27092b6c097edfa1f99d45bf972e28a86fbcc3dafd4"} Dec 06 16:00:21 crc kubenswrapper[4813]: I1206 16:00:21.567059 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26" event={"ID":"572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b","Type":"ContainerStarted","Data":"8c6a0c509163391b0d87264282452e25394b8d761cf70f3ede5d003e67bec3f5"} Dec 06 16:00:22 crc kubenswrapper[4813]: I1206 16:00:22.577643 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-rsbqf" event={"ID":"67c6b7ea-97cf-4ffa-9f3b-5111d4fbdeba","Type":"ContainerStarted","Data":"761468c3906a838356bd4651ded8e918193c750867da70ba38e553ab65d253c0"} Dec 06 16:00:23 crc kubenswrapper[4813]: I1206 16:00:23.590543 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26" event={"ID":"572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b","Type":"ContainerStarted","Data":"c7f2525fb901d862bda6d864a2e49cbf55f2edb6a21064ee525c9bacb3378a5f"} Dec 06 16:00:23 crc kubenswrapper[4813]: I1206 16:00:23.590984 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26" Dec 06 16:00:23 crc kubenswrapper[4813]: I1206 16:00:23.597503 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417280-bd6bw" event={"ID":"971a4ad6-5f62-47bc-b836-b7e6faaae668","Type":"ContainerStarted","Data":"e3f7658cda2aaf2a4a9df6e1927b6cf9c9670b5c65c1ad0d800a44fecf84f769"} Dec 06 16:00:23 crc kubenswrapper[4813]: I1206 16:00:23.610080 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-8zvf5" event={"ID":"5a8652b7-8a55-4658-acce-b7573421b94d","Type":"ContainerStarted","Data":"a1dff22f42fb241fed7e57d4cff82c46d1eb47500f1607eaf5d17bb4e5a975dd"} Dec 06 16:00:23 crc kubenswrapper[4813]: I1206 16:00:23.611617 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-cbmwf" event={"ID":"28acd95a-e657-46b5-a771-a24b27bdc6fe","Type":"ContainerStarted","Data":"c9faea236166e1d863d187beda1ee810d84e747db8ba643c0f0cdab8ba5b0934"} Dec 06 16:00:23 crc kubenswrapper[4813]: I1206 16:00:23.627871 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qkrqw" event={"ID":"6d54c577-3fdf-48af-9070-301dd3b3eb81","Type":"ContainerStarted","Data":"0befaeff8515914b20ad216b2fd560762da5e7b13471f301ef5be55bc1d6be42"} Dec 06 16:00:23 crc kubenswrapper[4813]: I1206 16:00:23.634408 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26" podStartSLOduration=38.634392553 podStartE2EDuration="38.634392553s" podCreationTimestamp="2025-12-06 15:59:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:00:23.62638126 +0000 UTC m=+863.517260846" watchObservedRunningTime="2025-12-06 16:00:23.634392553 +0000 UTC m=+863.525272129" Dec 06 16:00:23 crc kubenswrapper[4813]: E1206 16:00:23.745580 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-2mpwn" podUID="c35efae6-9bf9-466d-ac8d-203d8a2c4b97" Dec 06 16:00:23 crc kubenswrapper[4813]: E1206 16:00:23.908426 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8mt55" podUID="a8b0c607-9997-44e8-a3cc-a36e4b2e70d2" Dec 06 16:00:24 crc kubenswrapper[4813]: E1206 16:00:24.501035 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-wxsqx" podUID="e33ae276-0e3d-44b5-b970-17e98051c49c" Dec 06 16:00:24 crc kubenswrapper[4813]: I1206 16:00:24.640811 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-cxh72" event={"ID":"4d2f63b5-c93e-4949-bb79-459dad74dcf8","Type":"ContainerStarted","Data":"766961cf74534fda552a27c0ab8138d45fcd2e2ab66d1bb7f1eb267e879f922b"} Dec 06 16:00:24 crc kubenswrapper[4813]: I1206 16:00:24.644643 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-wxsqx" event={"ID":"e33ae276-0e3d-44b5-b970-17e98051c49c","Type":"ContainerStarted","Data":"8450ca63627e7f62b8b38005727aabf3459bdafd0eaaf2314dfef9e7ac9974a9"} Dec 06 16:00:24 crc kubenswrapper[4813]: I1206 16:00:24.650869 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-6rr2s" event={"ID":"b4cd1114-c317-4351-b566-05317dd589f1","Type":"ContainerStarted","Data":"60df93803293adc66d10abf2ee15432e8ed9863b547fa8b36675b7bd1c0f9047"} Dec 06 16:00:24 crc kubenswrapper[4813]: I1206 16:00:24.653738 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-g4ztj" event={"ID":"d67349e5-13fb-4859-ac8d-3e2c6bd67dbb","Type":"ContainerStarted","Data":"5704f2d02cd0d591fe57facdef55fa3bf52b99eaed67e5cc20a35c3cd1b5d9f2"} Dec 06 16:00:24 crc kubenswrapper[4813]: I1206 16:00:24.654839 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-cjzht" event={"ID":"8870fd51-2141-4cc8-956b-050e849dcbd4","Type":"ContainerStarted","Data":"896184cb82bc8ffe21ea1bced213604e41b23a9078421f8c9da7d2338cd573bf"} Dec 06 16:00:24 crc kubenswrapper[4813]: I1206 16:00:24.657389 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-8zvf5" event={"ID":"5a8652b7-8a55-4658-acce-b7573421b94d","Type":"ContainerStarted","Data":"68fb171fb966de5b1fb22973621688443c6fbbeb9fbc3b8c29a8baa941cea218"} Dec 06 16:00:24 crc kubenswrapper[4813]: I1206 16:00:24.657967 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-8zvf5" Dec 06 16:00:24 crc kubenswrapper[4813]: I1206 16:00:24.665351 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29417280-bd6bw" podStartSLOduration=24.665342056 podStartE2EDuration="24.665342056s" podCreationTimestamp="2025-12-06 16:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:00:23.660575225 +0000 UTC m=+863.551454801" watchObservedRunningTime="2025-12-06 16:00:24.665342056 +0000 UTC m=+864.556221632" Dec 06 16:00:24 crc kubenswrapper[4813]: I1206 16:00:24.665655 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qkrqw" event={"ID":"6d54c577-3fdf-48af-9070-301dd3b3eb81","Type":"ContainerStarted","Data":"54e1b158a7d8d4619f6054c66db819ed9d564d209f92ff3c755ba483ef65ed3f"} Dec 06 16:00:24 crc kubenswrapper[4813]: I1206 16:00:24.682043 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6dc7dc95b4-6wznz" event={"ID":"24d6dd0f-86e6-4b7a-beeb-56025a76ab5a","Type":"ContainerStarted","Data":"24518dcc0fd3438b7a231874e71137ff4ca40fe5785fdfc117c1461bd2d2dfa9"} Dec 06 16:00:24 crc kubenswrapper[4813]: I1206 16:00:24.691893 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-8zvf5" podStartSLOduration=3.367723803 podStartE2EDuration="39.691879367s" podCreationTimestamp="2025-12-06 15:59:45 +0000 UTC" firstStartedPulling="2025-12-06 15:59:47.19438551 +0000 UTC m=+827.085265086" lastFinishedPulling="2025-12-06 16:00:23.518541074 +0000 UTC m=+863.409420650" observedRunningTime="2025-12-06 16:00:24.688540053 +0000 UTC m=+864.579419629" watchObservedRunningTime="2025-12-06 16:00:24.691879367 +0000 UTC m=+864.582758943" Dec 06 16:00:24 crc kubenswrapper[4813]: I1206 16:00:24.699404 4813 generic.go:334] "Generic (PLEG): container finished" podID="971a4ad6-5f62-47bc-b836-b7e6faaae668" containerID="e3f7658cda2aaf2a4a9df6e1927b6cf9c9670b5c65c1ad0d800a44fecf84f769" exitCode=0 Dec 06 16:00:24 crc kubenswrapper[4813]: I1206 16:00:24.699546 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417280-bd6bw" event={"ID":"971a4ad6-5f62-47bc-b836-b7e6faaae668","Type":"ContainerDied","Data":"e3f7658cda2aaf2a4a9df6e1927b6cf9c9670b5c65c1ad0d800a44fecf84f769"} Dec 06 16:00:24 crc kubenswrapper[4813]: I1206 16:00:24.716383 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-2mpwn" event={"ID":"c35efae6-9bf9-466d-ac8d-203d8a2c4b97","Type":"ContainerStarted","Data":"83f8c1dbaf72ca697405c4784e063b07bb07fb18bd6b14f531317718bc0f7ad6"} Dec 06 16:00:24 crc kubenswrapper[4813]: I1206 16:00:24.726584 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-cbmwf" event={"ID":"28acd95a-e657-46b5-a771-a24b27bdc6fe","Type":"ContainerStarted","Data":"732e0f8d86f610911411bc0696880b1bbb4d1063501104f48e931eea311e3237"} Dec 06 16:00:24 crc kubenswrapper[4813]: I1206 16:00:24.726816 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-cbmwf" Dec 06 16:00:24 crc kubenswrapper[4813]: I1206 16:00:24.729143 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8mt55" event={"ID":"a8b0c607-9997-44e8-a3cc-a36e4b2e70d2","Type":"ContainerStarted","Data":"3a77283cef4b1a11ba72f941c1e8ae746d19bee24e83d88e2576289437833d1a"} Dec 06 16:00:24 crc kubenswrapper[4813]: I1206 16:00:24.732556 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-f9nfd" event={"ID":"3edd7ca8-a3ba-46c5-96c8-77aa9be672b7","Type":"ContainerStarted","Data":"5e5327003abba28f6013bfbbbd72774124e218076b15a7295cb6150fbc271154"} Dec 06 16:00:24 crc kubenswrapper[4813]: I1206 16:00:24.732579 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-f9nfd" Dec 06 16:00:24 crc kubenswrapper[4813]: I1206 16:00:24.769503 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-cbmwf" podStartSLOduration=7.382540013 podStartE2EDuration="39.769488369s" podCreationTimestamp="2025-12-06 15:59:45 +0000 UTC" firstStartedPulling="2025-12-06 15:59:47.21598774 +0000 UTC m=+827.106867316" lastFinishedPulling="2025-12-06 16:00:19.602936096 +0000 UTC m=+859.493815672" observedRunningTime="2025-12-06 16:00:24.76834586 +0000 UTC m=+864.659225446" watchObservedRunningTime="2025-12-06 16:00:24.769488369 +0000 UTC m=+864.660367945" Dec 06 16:00:24 crc kubenswrapper[4813]: I1206 16:00:24.820046 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-f9nfd" podStartSLOduration=3.847116971 podStartE2EDuration="40.820022657s" podCreationTimestamp="2025-12-06 15:59:44 +0000 UTC" firstStartedPulling="2025-12-06 15:59:46.570271711 +0000 UTC m=+826.461151287" lastFinishedPulling="2025-12-06 16:00:23.543177397 +0000 UTC m=+863.434056973" observedRunningTime="2025-12-06 16:00:24.815308797 +0000 UTC m=+864.706188373" watchObservedRunningTime="2025-12-06 16:00:24.820022657 +0000 UTC m=+864.710902223" Dec 06 16:00:25 crc kubenswrapper[4813]: I1206 16:00:25.245159 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-f9nfd" Dec 06 16:00:25 crc kubenswrapper[4813]: E1206 16:00:25.581657 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-g4ztj" podUID="d67349e5-13fb-4859-ac8d-3e2c6bd67dbb" Dec 06 16:00:25 crc kubenswrapper[4813]: I1206 16:00:25.763427 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qkrqw" podStartSLOduration=8.284043332 podStartE2EDuration="40.763401956s" podCreationTimestamp="2025-12-06 15:59:45 +0000 UTC" firstStartedPulling="2025-12-06 15:59:47.124516745 +0000 UTC m=+827.015396321" lastFinishedPulling="2025-12-06 16:00:19.603875359 +0000 UTC m=+859.494754945" observedRunningTime="2025-12-06 16:00:25.761549919 +0000 UTC m=+865.652429505" watchObservedRunningTime="2025-12-06 16:00:25.763401956 +0000 UTC m=+865.654281572" Dec 06 16:00:25 crc kubenswrapper[4813]: E1206 16:00:25.828921 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-cjzht" podUID="8870fd51-2141-4cc8-956b-050e849dcbd4" Dec 06 16:00:25 crc kubenswrapper[4813]: E1206 16:00:25.860924 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-6dc7dc95b4-6wznz" podUID="24d6dd0f-86e6-4b7a-beeb-56025a76ab5a" Dec 06 16:00:26 crc kubenswrapper[4813]: E1206 16:00:26.092741 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6n2hj" podUID="99a0eae7-c990-42cd-a18e-78b81177bbb6" Dec 06 16:00:26 crc kubenswrapper[4813]: E1206 16:00:26.095896 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-78fv6" podUID="b675acf0-51cc-43d9-b9c6-bd0ee1730d25" Dec 06 16:00:26 crc kubenswrapper[4813]: I1206 16:00:26.136867 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qkrqw" Dec 06 16:00:26 crc kubenswrapper[4813]: I1206 16:00:26.752680 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-78fv6" event={"ID":"b675acf0-51cc-43d9-b9c6-bd0ee1730d25","Type":"ContainerStarted","Data":"508802e49e8569bfba5dab29ae45c571c4036e72ecabd2aa737b2cbc9d76c665"} Dec 06 16:00:26 crc kubenswrapper[4813]: I1206 16:00:26.756820 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6n2hj" event={"ID":"99a0eae7-c990-42cd-a18e-78b81177bbb6","Type":"ContainerStarted","Data":"fad92c0b46414a2e22d5a7c01b8bedeffb97df1c3a861258532d78ecba4201cb"} Dec 06 16:00:26 crc kubenswrapper[4813]: I1206 16:00:26.763009 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-g9q96" event={"ID":"031a9bd1-298a-4945-a52d-452005c8467e","Type":"ContainerStarted","Data":"2dfcd97fa82c61f099eea60cd5c532413e0216e79bcf0d90e3b08320701f6204"} Dec 06 16:00:26 crc kubenswrapper[4813]: I1206 16:00:26.788553 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-967d97867-cxh72" podStartSLOduration=5.99767483 podStartE2EDuration="42.788528061s" podCreationTimestamp="2025-12-06 15:59:44 +0000 UTC" firstStartedPulling="2025-12-06 15:59:46.936707307 +0000 UTC m=+826.827586883" lastFinishedPulling="2025-12-06 16:00:23.727560548 +0000 UTC m=+863.618440114" observedRunningTime="2025-12-06 16:00:26.781996376 +0000 UTC m=+866.672875962" watchObservedRunningTime="2025-12-06 16:00:26.788528061 +0000 UTC m=+866.679407667" Dec 06 16:00:26 crc kubenswrapper[4813]: I1206 16:00:26.836206 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-g9q96" podStartSLOduration=4.838048035 podStartE2EDuration="41.836180166s" podCreationTimestamp="2025-12-06 15:59:45 +0000 UTC" firstStartedPulling="2025-12-06 15:59:47.035342411 +0000 UTC m=+826.926221987" lastFinishedPulling="2025-12-06 16:00:24.033474542 +0000 UTC m=+863.924354118" observedRunningTime="2025-12-06 16:00:26.826826969 +0000 UTC m=+866.717706565" watchObservedRunningTime="2025-12-06 16:00:26.836180166 +0000 UTC m=+866.727059752" Dec 06 16:00:26 crc kubenswrapper[4813]: I1206 16:00:26.857838 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-6rr2s" podStartSLOduration=4.764721348 podStartE2EDuration="41.857815513s" podCreationTimestamp="2025-12-06 15:59:45 +0000 UTC" firstStartedPulling="2025-12-06 15:59:46.783290616 +0000 UTC m=+826.674170192" lastFinishedPulling="2025-12-06 16:00:23.876384781 +0000 UTC m=+863.767264357" observedRunningTime="2025-12-06 16:00:26.852311933 +0000 UTC m=+866.743191519" watchObservedRunningTime="2025-12-06 16:00:26.857815513 +0000 UTC m=+866.748695089" Dec 06 16:00:27 crc kubenswrapper[4813]: I1206 16:00:27.456183 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417280-bd6bw" Dec 06 16:00:27 crc kubenswrapper[4813]: I1206 16:00:27.525842 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/971a4ad6-5f62-47bc-b836-b7e6faaae668-secret-volume\") pod \"971a4ad6-5f62-47bc-b836-b7e6faaae668\" (UID: \"971a4ad6-5f62-47bc-b836-b7e6faaae668\") " Dec 06 16:00:27 crc kubenswrapper[4813]: I1206 16:00:27.525916 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fjdk\" (UniqueName: \"kubernetes.io/projected/971a4ad6-5f62-47bc-b836-b7e6faaae668-kube-api-access-8fjdk\") pod \"971a4ad6-5f62-47bc-b836-b7e6faaae668\" (UID: \"971a4ad6-5f62-47bc-b836-b7e6faaae668\") " Dec 06 16:00:27 crc kubenswrapper[4813]: I1206 16:00:27.526967 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/971a4ad6-5f62-47bc-b836-b7e6faaae668-config-volume\") pod \"971a4ad6-5f62-47bc-b836-b7e6faaae668\" (UID: \"971a4ad6-5f62-47bc-b836-b7e6faaae668\") " Dec 06 16:00:27 crc kubenswrapper[4813]: I1206 16:00:27.527558 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/971a4ad6-5f62-47bc-b836-b7e6faaae668-config-volume" (OuterVolumeSpecName: "config-volume") pod "971a4ad6-5f62-47bc-b836-b7e6faaae668" (UID: "971a4ad6-5f62-47bc-b836-b7e6faaae668"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:00:27 crc kubenswrapper[4813]: I1206 16:00:27.534894 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/971a4ad6-5f62-47bc-b836-b7e6faaae668-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "971a4ad6-5f62-47bc-b836-b7e6faaae668" (UID: "971a4ad6-5f62-47bc-b836-b7e6faaae668"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:00:27 crc kubenswrapper[4813]: I1206 16:00:27.534924 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/971a4ad6-5f62-47bc-b836-b7e6faaae668-kube-api-access-8fjdk" (OuterVolumeSpecName: "kube-api-access-8fjdk") pod "971a4ad6-5f62-47bc-b836-b7e6faaae668" (UID: "971a4ad6-5f62-47bc-b836-b7e6faaae668"). InnerVolumeSpecName "kube-api-access-8fjdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:00:27 crc kubenswrapper[4813]: I1206 16:00:27.629182 4813 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/971a4ad6-5f62-47bc-b836-b7e6faaae668-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 16:00:27 crc kubenswrapper[4813]: I1206 16:00:27.629230 4813 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/971a4ad6-5f62-47bc-b836-b7e6faaae668-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 16:00:27 crc kubenswrapper[4813]: I1206 16:00:27.629270 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fjdk\" (UniqueName: \"kubernetes.io/projected/971a4ad6-5f62-47bc-b836-b7e6faaae668-kube-api-access-8fjdk\") on node \"crc\" DevicePath \"\"" Dec 06 16:00:27 crc kubenswrapper[4813]: I1206 16:00:27.777782 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417280-bd6bw" event={"ID":"971a4ad6-5f62-47bc-b836-b7e6faaae668","Type":"ContainerDied","Data":"60aa677834ba1c5954e002df4b024349e8a0c50a3b67e634a5102ff1235e5fa3"} Dec 06 16:00:27 crc kubenswrapper[4813]: I1206 16:00:27.778072 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="60aa677834ba1c5954e002df4b024349e8a0c50a3b67e634a5102ff1235e5fa3" Dec 06 16:00:27 crc kubenswrapper[4813]: I1206 16:00:27.778096 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-g9q96" Dec 06 16:00:27 crc kubenswrapper[4813]: I1206 16:00:27.777803 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417280-bd6bw" Dec 06 16:00:27 crc kubenswrapper[4813]: I1206 16:00:27.780239 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-g9q96" Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.115020 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-5c8695d769-78g26" Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.787503 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6dc7dc95b4-6wznz" event={"ID":"24d6dd0f-86e6-4b7a-beeb-56025a76ab5a","Type":"ContainerStarted","Data":"9694d21b8c829bf664c92eec3ed2f3da1a6fceae35fdfda26ad1aabac637fa81"} Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.787899 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-6dc7dc95b4-6wznz" Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.789276 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-x8vkl" event={"ID":"b05dafaa-0d81-4b80-ad3e-48da48a1655f","Type":"ContainerStarted","Data":"a713543b0f25c504222721305f97ec270a59a85d855ddbe6c526542291bff274"} Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.789457 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-x8vkl" Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.790561 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-6p6sf" event={"ID":"a8763d41-9404-43ec-866b-9244bf2c4af9","Type":"ContainerStarted","Data":"08d74b0caf0bf55e5bd595da80f51e51c193bd0386b16a62f0c7ba1d2dfd9a67"} Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.791199 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-6p6sf" Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.791529 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-x8vkl" Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.792446 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-g4ztj" event={"ID":"d67349e5-13fb-4859-ac8d-3e2c6bd67dbb","Type":"ContainerStarted","Data":"d658f78e8dea160b4fe463a69c31ffcf8fa299e768d607c0612f0c56fea51d53"} Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.792623 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-g4ztj" Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.793818 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-rsbqf" event={"ID":"67c6b7ea-97cf-4ffa-9f3b-5111d4fbdeba","Type":"ContainerStarted","Data":"295d88a9114c0db2390c136e43050e9cdfb672e88687657111b0c9ec18602db2"} Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.793966 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-rsbqf" Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.796216 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6n2hj" event={"ID":"99a0eae7-c990-42cd-a18e-78b81177bbb6","Type":"ContainerStarted","Data":"4b20f79954a10d4a636b9094fb25fe9810bdfebf029a6b0a7b975d4f060e3643"} Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.796301 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6n2hj" Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.796646 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-6p6sf" Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.797513 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-4rdzx" event={"ID":"218be173-f41f-48b1-8af0-b94bbc483e8b","Type":"ContainerStarted","Data":"b50903b1a61014a81c7e94e0264714f09af0001daf80b2e085c352fa1c73fcb2"} Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.797684 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-4rdzx" Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.801201 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-78fv6" event={"ID":"b675acf0-51cc-43d9-b9c6-bd0ee1730d25","Type":"ContainerStarted","Data":"6e95e0fecc88826482702d4f98d3ec4ba8c3a1f50ecd31bfc0ea3a481cb682e2"} Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.801742 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-rsbqf" Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.801780 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-78fv6" Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.802810 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-4rdzx" Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.803680 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8mt55" event={"ID":"a8b0c607-9997-44e8-a3cc-a36e4b2e70d2","Type":"ContainerStarted","Data":"722ed44a3abcc15ce578bf576e3976e835c0735f2835655ce04fc09648892826"} Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.804078 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8mt55" Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.805331 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fc7jrk" event={"ID":"279a2707-03fd-4cbc-9e13-d7a0d2a284db","Type":"ContainerStarted","Data":"3639cd08f9cae3c3c30931ca4eebc6b41f2cdbf631f2389cd4dacc729bbe18f7"} Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.805355 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fc7jrk" event={"ID":"279a2707-03fd-4cbc-9e13-d7a0d2a284db","Type":"ContainerStarted","Data":"dc8b523413992f037bb455a65d602d252558f98cc834b2f7b62510f04f40ee69"} Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.805700 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fc7jrk" Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.807102 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-2mpwn" event={"ID":"c35efae6-9bf9-466d-ac8d-203d8a2c4b97","Type":"ContainerStarted","Data":"43205d6ed5de3367c34b61d50c0862911f7fa2e5ecd0c654b6aab63eeb891a45"} Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.807212 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-2mpwn" Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.809039 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-wxsqx" event={"ID":"e33ae276-0e3d-44b5-b970-17e98051c49c","Type":"ContainerStarted","Data":"fa21eb9c2ce8c9de3256086d307daac91a86ca424cbd7ce30f41c27786892943"} Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.809175 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-wxsqx" Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.810732 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pkvxj" event={"ID":"6983648f-60ef-429d-a2fd-de5dc7e6d0ba","Type":"ContainerStarted","Data":"07c434a99683f47a9c4f29270e009260cf88136ef4767bfd355cb9d1b6964496"} Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.810764 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pkvxj" event={"ID":"6983648f-60ef-429d-a2fd-de5dc7e6d0ba","Type":"ContainerStarted","Data":"6cf60be3b5d218122cbc3d26b1404a62969b9361b7dd4a9623f1f4ef60804e7b"} Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.810868 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pkvxj" Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.812359 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-cjzht" event={"ID":"8870fd51-2141-4cc8-956b-050e849dcbd4","Type":"ContainerStarted","Data":"c6334ecaf41b7fff80049714aee34b78367937e8c477997124aef762bb7a3464"} Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.812618 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-cjzht" Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.831875 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-6dc7dc95b4-6wznz" podStartSLOduration=3.710686254 podStartE2EDuration="44.831853788s" podCreationTimestamp="2025-12-06 15:59:44 +0000 UTC" firstStartedPulling="2025-12-06 15:59:46.571787681 +0000 UTC m=+826.462667257" lastFinishedPulling="2025-12-06 16:00:27.692955215 +0000 UTC m=+867.583834791" observedRunningTime="2025-12-06 16:00:28.825012435 +0000 UTC m=+868.715892011" watchObservedRunningTime="2025-12-06 16:00:28.831853788 +0000 UTC m=+868.722733364" Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.856476 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8mt55" podStartSLOduration=3.779093191 podStartE2EDuration="44.85646058s" podCreationTimestamp="2025-12-06 15:59:44 +0000 UTC" firstStartedPulling="2025-12-06 15:59:46.614626232 +0000 UTC m=+826.505505798" lastFinishedPulling="2025-12-06 16:00:27.691993611 +0000 UTC m=+867.582873187" observedRunningTime="2025-12-06 16:00:28.853276539 +0000 UTC m=+868.744156115" watchObservedRunningTime="2025-12-06 16:00:28.85646058 +0000 UTC m=+868.747340156" Dec 06 16:00:28 crc kubenswrapper[4813]: I1206 16:00:28.880061 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-rsbqf" podStartSLOduration=3.325620285 podStartE2EDuration="43.880043306s" podCreationTimestamp="2025-12-06 15:59:45 +0000 UTC" firstStartedPulling="2025-12-06 15:59:47.129342202 +0000 UTC m=+827.020221778" lastFinishedPulling="2025-12-06 16:00:27.683765223 +0000 UTC m=+867.574644799" observedRunningTime="2025-12-06 16:00:28.879680497 +0000 UTC m=+868.770560073" watchObservedRunningTime="2025-12-06 16:00:28.880043306 +0000 UTC m=+868.770922882" Dec 06 16:00:29 crc kubenswrapper[4813]: I1206 16:00:29.046189 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-4rdzx" podStartSLOduration=3.7347414519999997 podStartE2EDuration="45.046175986s" podCreationTimestamp="2025-12-06 15:59:44 +0000 UTC" firstStartedPulling="2025-12-06 15:59:46.145363071 +0000 UTC m=+826.036242647" lastFinishedPulling="2025-12-06 16:00:27.456797595 +0000 UTC m=+867.347677181" observedRunningTime="2025-12-06 16:00:28.963391723 +0000 UTC m=+868.854271299" watchObservedRunningTime="2025-12-06 16:00:29.046175986 +0000 UTC m=+868.937055562" Dec 06 16:00:29 crc kubenswrapper[4813]: I1206 16:00:29.050035 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-cjzht" podStartSLOduration=4.130155855 podStartE2EDuration="45.050028983s" podCreationTimestamp="2025-12-06 15:59:44 +0000 UTC" firstStartedPulling="2025-12-06 15:59:46.782951427 +0000 UTC m=+826.673831003" lastFinishedPulling="2025-12-06 16:00:27.702824555 +0000 UTC m=+867.593704131" observedRunningTime="2025-12-06 16:00:29.043628372 +0000 UTC m=+868.934507948" watchObservedRunningTime="2025-12-06 16:00:29.050028983 +0000 UTC m=+868.940908559" Dec 06 16:00:29 crc kubenswrapper[4813]: I1206 16:00:29.069030 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-g4ztj" podStartSLOduration=3.819977441 podStartE2EDuration="45.069015723s" podCreationTimestamp="2025-12-06 15:59:44 +0000 UTC" firstStartedPulling="2025-12-06 15:59:46.443810201 +0000 UTC m=+826.334689777" lastFinishedPulling="2025-12-06 16:00:27.692848473 +0000 UTC m=+867.583728059" observedRunningTime="2025-12-06 16:00:29.064712735 +0000 UTC m=+868.955592311" watchObservedRunningTime="2025-12-06 16:00:29.069015723 +0000 UTC m=+868.959895299" Dec 06 16:00:29 crc kubenswrapper[4813]: I1206 16:00:29.114636 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-x8vkl" podStartSLOduration=4.5393738500000005 podStartE2EDuration="44.114620176s" podCreationTimestamp="2025-12-06 15:59:45 +0000 UTC" firstStartedPulling="2025-12-06 15:59:47.129655571 +0000 UTC m=+827.020535147" lastFinishedPulling="2025-12-06 16:00:26.704901897 +0000 UTC m=+866.595781473" observedRunningTime="2025-12-06 16:00:29.103879715 +0000 UTC m=+868.994759301" watchObservedRunningTime="2025-12-06 16:00:29.114620176 +0000 UTC m=+869.005499752" Dec 06 16:00:29 crc kubenswrapper[4813]: I1206 16:00:29.153140 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-78fv6" podStartSLOduration=3.548368704 podStartE2EDuration="44.15312664s" podCreationTimestamp="2025-12-06 15:59:45 +0000 UTC" firstStartedPulling="2025-12-06 15:59:47.072968755 +0000 UTC m=+826.963848331" lastFinishedPulling="2025-12-06 16:00:27.677726691 +0000 UTC m=+867.568606267" observedRunningTime="2025-12-06 16:00:29.152375561 +0000 UTC m=+869.043255137" watchObservedRunningTime="2025-12-06 16:00:29.15312664 +0000 UTC m=+869.044006206" Dec 06 16:00:29 crc kubenswrapper[4813]: I1206 16:00:29.185006 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pkvxj" podStartSLOduration=38.263426465 podStartE2EDuration="45.184986755s" podCreationTimestamp="2025-12-06 15:59:44 +0000 UTC" firstStartedPulling="2025-12-06 16:00:20.737080558 +0000 UTC m=+860.627960134" lastFinishedPulling="2025-12-06 16:00:27.658640848 +0000 UTC m=+867.549520424" observedRunningTime="2025-12-06 16:00:29.181168889 +0000 UTC m=+869.072048465" watchObservedRunningTime="2025-12-06 16:00:29.184986755 +0000 UTC m=+869.075866331" Dec 06 16:00:29 crc kubenswrapper[4813]: I1206 16:00:29.203612 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-wxsqx" podStartSLOduration=4.022754423 podStartE2EDuration="45.203595256s" podCreationTimestamp="2025-12-06 15:59:44 +0000 UTC" firstStartedPulling="2025-12-06 15:59:46.483218912 +0000 UTC m=+826.374098488" lastFinishedPulling="2025-12-06 16:00:27.664059745 +0000 UTC m=+867.554939321" observedRunningTime="2025-12-06 16:00:29.202774765 +0000 UTC m=+869.093654341" watchObservedRunningTime="2025-12-06 16:00:29.203595256 +0000 UTC m=+869.094474822" Dec 06 16:00:29 crc kubenswrapper[4813]: I1206 16:00:29.275191 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-6p6sf" podStartSLOduration=4.371167599 podStartE2EDuration="44.275169395s" podCreationTimestamp="2025-12-06 15:59:45 +0000 UTC" firstStartedPulling="2025-12-06 15:59:46.801021754 +0000 UTC m=+826.691901320" lastFinishedPulling="2025-12-06 16:00:26.70502354 +0000 UTC m=+866.595903116" observedRunningTime="2025-12-06 16:00:29.220034101 +0000 UTC m=+869.110913677" watchObservedRunningTime="2025-12-06 16:00:29.275169395 +0000 UTC m=+869.166048991" Dec 06 16:00:29 crc kubenswrapper[4813]: I1206 16:00:29.306457 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fc7jrk" podStartSLOduration=37.356161239 podStartE2EDuration="44.306436706s" podCreationTimestamp="2025-12-06 15:59:45 +0000 UTC" firstStartedPulling="2025-12-06 16:00:20.718435776 +0000 UTC m=+860.609315352" lastFinishedPulling="2025-12-06 16:00:27.668711243 +0000 UTC m=+867.559590819" observedRunningTime="2025-12-06 16:00:29.292338799 +0000 UTC m=+869.183218445" watchObservedRunningTime="2025-12-06 16:00:29.306436706 +0000 UTC m=+869.197316282" Dec 06 16:00:29 crc kubenswrapper[4813]: I1206 16:00:29.331089 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-2mpwn" podStartSLOduration=3.711880518 podStartE2EDuration="44.331072918s" podCreationTimestamp="2025-12-06 15:59:45 +0000 UTC" firstStartedPulling="2025-12-06 15:59:47.072843572 +0000 UTC m=+826.963723148" lastFinishedPulling="2025-12-06 16:00:27.692035972 +0000 UTC m=+867.582915548" observedRunningTime="2025-12-06 16:00:29.329442927 +0000 UTC m=+869.220322503" watchObservedRunningTime="2025-12-06 16:00:29.331072918 +0000 UTC m=+869.221952494" Dec 06 16:00:29 crc kubenswrapper[4813]: I1206 16:00:29.381053 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6n2hj" podStartSLOduration=3.766111793 podStartE2EDuration="44.381037562s" podCreationTimestamp="2025-12-06 15:59:45 +0000 UTC" firstStartedPulling="2025-12-06 15:59:47.076696043 +0000 UTC m=+826.967575619" lastFinishedPulling="2025-12-06 16:00:27.691621812 +0000 UTC m=+867.582501388" observedRunningTime="2025-12-06 16:00:29.375498102 +0000 UTC m=+869.266377678" watchObservedRunningTime="2025-12-06 16:00:29.381037562 +0000 UTC m=+869.271917138" Dec 06 16:00:30 crc kubenswrapper[4813]: I1206 16:00:30.244123 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ckh76"] Dec 06 16:00:30 crc kubenswrapper[4813]: E1206 16:00:30.244712 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="971a4ad6-5f62-47bc-b836-b7e6faaae668" containerName="collect-profiles" Dec 06 16:00:30 crc kubenswrapper[4813]: I1206 16:00:30.244724 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="971a4ad6-5f62-47bc-b836-b7e6faaae668" containerName="collect-profiles" Dec 06 16:00:30 crc kubenswrapper[4813]: I1206 16:00:30.245347 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="971a4ad6-5f62-47bc-b836-b7e6faaae668" containerName="collect-profiles" Dec 06 16:00:30 crc kubenswrapper[4813]: I1206 16:00:30.246453 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ckh76" Dec 06 16:00:30 crc kubenswrapper[4813]: I1206 16:00:30.262473 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ckh76"] Dec 06 16:00:30 crc kubenswrapper[4813]: I1206 16:00:30.386878 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/788502af-b07e-47a4-ad30-3b0181a8104c-utilities\") pod \"community-operators-ckh76\" (UID: \"788502af-b07e-47a4-ad30-3b0181a8104c\") " pod="openshift-marketplace/community-operators-ckh76" Dec 06 16:00:30 crc kubenswrapper[4813]: I1206 16:00:30.386934 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/788502af-b07e-47a4-ad30-3b0181a8104c-catalog-content\") pod \"community-operators-ckh76\" (UID: \"788502af-b07e-47a4-ad30-3b0181a8104c\") " pod="openshift-marketplace/community-operators-ckh76" Dec 06 16:00:30 crc kubenswrapper[4813]: I1206 16:00:30.387024 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x66j2\" (UniqueName: \"kubernetes.io/projected/788502af-b07e-47a4-ad30-3b0181a8104c-kube-api-access-x66j2\") pod \"community-operators-ckh76\" (UID: \"788502af-b07e-47a4-ad30-3b0181a8104c\") " pod="openshift-marketplace/community-operators-ckh76" Dec 06 16:00:30 crc kubenswrapper[4813]: I1206 16:00:30.491056 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x66j2\" (UniqueName: \"kubernetes.io/projected/788502af-b07e-47a4-ad30-3b0181a8104c-kube-api-access-x66j2\") pod \"community-operators-ckh76\" (UID: \"788502af-b07e-47a4-ad30-3b0181a8104c\") " pod="openshift-marketplace/community-operators-ckh76" Dec 06 16:00:30 crc kubenswrapper[4813]: I1206 16:00:30.491134 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/788502af-b07e-47a4-ad30-3b0181a8104c-utilities\") pod \"community-operators-ckh76\" (UID: \"788502af-b07e-47a4-ad30-3b0181a8104c\") " pod="openshift-marketplace/community-operators-ckh76" Dec 06 16:00:30 crc kubenswrapper[4813]: I1206 16:00:30.491162 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/788502af-b07e-47a4-ad30-3b0181a8104c-catalog-content\") pod \"community-operators-ckh76\" (UID: \"788502af-b07e-47a4-ad30-3b0181a8104c\") " pod="openshift-marketplace/community-operators-ckh76" Dec 06 16:00:30 crc kubenswrapper[4813]: I1206 16:00:30.491566 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/788502af-b07e-47a4-ad30-3b0181a8104c-catalog-content\") pod \"community-operators-ckh76\" (UID: \"788502af-b07e-47a4-ad30-3b0181a8104c\") " pod="openshift-marketplace/community-operators-ckh76" Dec 06 16:00:30 crc kubenswrapper[4813]: I1206 16:00:30.491986 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/788502af-b07e-47a4-ad30-3b0181a8104c-utilities\") pod \"community-operators-ckh76\" (UID: \"788502af-b07e-47a4-ad30-3b0181a8104c\") " pod="openshift-marketplace/community-operators-ckh76" Dec 06 16:00:30 crc kubenswrapper[4813]: I1206 16:00:30.513412 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x66j2\" (UniqueName: \"kubernetes.io/projected/788502af-b07e-47a4-ad30-3b0181a8104c-kube-api-access-x66j2\") pod \"community-operators-ckh76\" (UID: \"788502af-b07e-47a4-ad30-3b0181a8104c\") " pod="openshift-marketplace/community-operators-ckh76" Dec 06 16:00:30 crc kubenswrapper[4813]: I1206 16:00:30.567766 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ckh76" Dec 06 16:00:30 crc kubenswrapper[4813]: I1206 16:00:30.847654 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9lwhh" event={"ID":"00cfc20a-68d6-46be-98b8-83a43c9d11d9","Type":"ContainerStarted","Data":"8fa3ed44b29b5cc4c6c0034322d2bb906a773f7e5fc6fdf8986dfce59910e659"} Dec 06 16:00:30 crc kubenswrapper[4813]: I1206 16:00:30.868820 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9lwhh" podStartSLOduration=3.14979339 podStartE2EDuration="45.868806563s" podCreationTimestamp="2025-12-06 15:59:45 +0000 UTC" firstStartedPulling="2025-12-06 15:59:47.289968264 +0000 UTC m=+827.180847840" lastFinishedPulling="2025-12-06 16:00:30.008981437 +0000 UTC m=+869.899861013" observedRunningTime="2025-12-06 16:00:30.864705639 +0000 UTC m=+870.755585215" watchObservedRunningTime="2025-12-06 16:00:30.868806563 +0000 UTC m=+870.759686139" Dec 06 16:00:30 crc kubenswrapper[4813]: I1206 16:00:30.934421 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ckh76"] Dec 06 16:00:31 crc kubenswrapper[4813]: I1206 16:00:31.859065 4813 generic.go:334] "Generic (PLEG): container finished" podID="788502af-b07e-47a4-ad30-3b0181a8104c" containerID="1d208ab8d69a9ba74e68ee73f34291e4a7c37705890ab63473b07eed784332b0" exitCode=0 Dec 06 16:00:31 crc kubenswrapper[4813]: I1206 16:00:31.859312 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ckh76" event={"ID":"788502af-b07e-47a4-ad30-3b0181a8104c","Type":"ContainerDied","Data":"1d208ab8d69a9ba74e68ee73f34291e4a7c37705890ab63473b07eed784332b0"} Dec 06 16:00:31 crc kubenswrapper[4813]: I1206 16:00:31.859488 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ckh76" event={"ID":"788502af-b07e-47a4-ad30-3b0181a8104c","Type":"ContainerStarted","Data":"a63ec2127060213fbf39e0e5c5fbb9d34e6d52cd0b809739f0904d24beb1d69f"} Dec 06 16:00:32 crc kubenswrapper[4813]: I1206 16:00:32.877387 4813 generic.go:334] "Generic (PLEG): container finished" podID="788502af-b07e-47a4-ad30-3b0181a8104c" containerID="9ac75f149df2eea2bf3748f138d5bc85a05cbb035e8300e01183949956613b67" exitCode=0 Dec 06 16:00:32 crc kubenswrapper[4813]: I1206 16:00:32.877781 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ckh76" event={"ID":"788502af-b07e-47a4-ad30-3b0181a8104c","Type":"ContainerDied","Data":"9ac75f149df2eea2bf3748f138d5bc85a05cbb035e8300e01183949956613b67"} Dec 06 16:00:34 crc kubenswrapper[4813]: I1206 16:00:34.911588 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ckh76" event={"ID":"788502af-b07e-47a4-ad30-3b0181a8104c","Type":"ContainerStarted","Data":"e0c40a65e0b77b681e5c4eababbfe34a203943873bb0c4b3a29fac53c9be75a0"} Dec 06 16:00:34 crc kubenswrapper[4813]: I1206 16:00:34.947245 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ckh76" podStartSLOduration=2.204408498 podStartE2EDuration="4.947226127s" podCreationTimestamp="2025-12-06 16:00:30 +0000 UTC" firstStartedPulling="2025-12-06 16:00:31.861973501 +0000 UTC m=+871.752853107" lastFinishedPulling="2025-12-06 16:00:34.60479113 +0000 UTC m=+874.495670736" observedRunningTime="2025-12-06 16:00:34.938556698 +0000 UTC m=+874.829436314" watchObservedRunningTime="2025-12-06 16:00:34.947226127 +0000 UTC m=+874.838105713" Dec 06 16:00:35 crc kubenswrapper[4813]: I1206 16:00:35.101088 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-6dc7dc95b4-6wznz" Dec 06 16:00:35 crc kubenswrapper[4813]: I1206 16:00:35.137032 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8mt55" Dec 06 16:00:35 crc kubenswrapper[4813]: I1206 16:00:35.298822 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-wxsqx" Dec 06 16:00:35 crc kubenswrapper[4813]: I1206 16:00:35.339234 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-g4ztj" Dec 06 16:00:35 crc kubenswrapper[4813]: I1206 16:00:35.457496 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-967d97867-cxh72" Dec 06 16:00:35 crc kubenswrapper[4813]: I1206 16:00:35.460136 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-967d97867-cxh72" Dec 06 16:00:35 crc kubenswrapper[4813]: I1206 16:00:35.512493 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-cjzht" Dec 06 16:00:35 crc kubenswrapper[4813]: I1206 16:00:35.592956 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6n2hj" Dec 06 16:00:35 crc kubenswrapper[4813]: I1206 16:00:35.646058 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-6rr2s" Dec 06 16:00:35 crc kubenswrapper[4813]: I1206 16:00:35.648811 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-6rr2s" Dec 06 16:00:35 crc kubenswrapper[4813]: I1206 16:00:35.694866 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-78fv6" Dec 06 16:00:35 crc kubenswrapper[4813]: I1206 16:00:35.937748 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-2mpwn" Dec 06 16:00:36 crc kubenswrapper[4813]: I1206 16:00:36.139705 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-qkrqw" Dec 06 16:00:36 crc kubenswrapper[4813]: I1206 16:00:36.201633 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-cbmwf" Dec 06 16:00:36 crc kubenswrapper[4813]: I1206 16:00:36.316847 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-8zvf5" Dec 06 16:00:37 crc kubenswrapper[4813]: I1206 16:00:37.214830 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pkvxj" Dec 06 16:00:37 crc kubenswrapper[4813]: I1206 16:00:37.677370 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fc7jrk" Dec 06 16:00:40 crc kubenswrapper[4813]: I1206 16:00:40.568493 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ckh76" Dec 06 16:00:40 crc kubenswrapper[4813]: I1206 16:00:40.568895 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ckh76" Dec 06 16:00:40 crc kubenswrapper[4813]: I1206 16:00:40.705780 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ckh76" Dec 06 16:00:40 crc kubenswrapper[4813]: I1206 16:00:40.992835 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ckh76" Dec 06 16:00:41 crc kubenswrapper[4813]: I1206 16:00:41.050859 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ckh76"] Dec 06 16:00:42 crc kubenswrapper[4813]: I1206 16:00:42.966250 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ckh76" podUID="788502af-b07e-47a4-ad30-3b0181a8104c" containerName="registry-server" containerID="cri-o://e0c40a65e0b77b681e5c4eababbfe34a203943873bb0c4b3a29fac53c9be75a0" gracePeriod=2 Dec 06 16:00:43 crc kubenswrapper[4813]: I1206 16:00:43.359323 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-g8tcs"] Dec 06 16:00:43 crc kubenswrapper[4813]: I1206 16:00:43.362554 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8tcs" Dec 06 16:00:43 crc kubenswrapper[4813]: I1206 16:00:43.380743 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g8tcs"] Dec 06 16:00:43 crc kubenswrapper[4813]: I1206 16:00:43.453672 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ckh76" Dec 06 16:00:43 crc kubenswrapper[4813]: I1206 16:00:43.533420 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9cd8622-0e78-400c-9de3-213fe41d1c1a-catalog-content\") pod \"redhat-operators-g8tcs\" (UID: \"e9cd8622-0e78-400c-9de3-213fe41d1c1a\") " pod="openshift-marketplace/redhat-operators-g8tcs" Dec 06 16:00:43 crc kubenswrapper[4813]: I1206 16:00:43.533493 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9vgp\" (UniqueName: \"kubernetes.io/projected/e9cd8622-0e78-400c-9de3-213fe41d1c1a-kube-api-access-n9vgp\") pod \"redhat-operators-g8tcs\" (UID: \"e9cd8622-0e78-400c-9de3-213fe41d1c1a\") " pod="openshift-marketplace/redhat-operators-g8tcs" Dec 06 16:00:43 crc kubenswrapper[4813]: I1206 16:00:43.533531 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9cd8622-0e78-400c-9de3-213fe41d1c1a-utilities\") pod \"redhat-operators-g8tcs\" (UID: \"e9cd8622-0e78-400c-9de3-213fe41d1c1a\") " pod="openshift-marketplace/redhat-operators-g8tcs" Dec 06 16:00:43 crc kubenswrapper[4813]: I1206 16:00:43.634133 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x66j2\" (UniqueName: \"kubernetes.io/projected/788502af-b07e-47a4-ad30-3b0181a8104c-kube-api-access-x66j2\") pod \"788502af-b07e-47a4-ad30-3b0181a8104c\" (UID: \"788502af-b07e-47a4-ad30-3b0181a8104c\") " Dec 06 16:00:43 crc kubenswrapper[4813]: I1206 16:00:43.634187 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/788502af-b07e-47a4-ad30-3b0181a8104c-catalog-content\") pod \"788502af-b07e-47a4-ad30-3b0181a8104c\" (UID: \"788502af-b07e-47a4-ad30-3b0181a8104c\") " Dec 06 16:00:43 crc kubenswrapper[4813]: I1206 16:00:43.634317 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/788502af-b07e-47a4-ad30-3b0181a8104c-utilities\") pod \"788502af-b07e-47a4-ad30-3b0181a8104c\" (UID: \"788502af-b07e-47a4-ad30-3b0181a8104c\") " Dec 06 16:00:43 crc kubenswrapper[4813]: I1206 16:00:43.634453 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9cd8622-0e78-400c-9de3-213fe41d1c1a-catalog-content\") pod \"redhat-operators-g8tcs\" (UID: \"e9cd8622-0e78-400c-9de3-213fe41d1c1a\") " pod="openshift-marketplace/redhat-operators-g8tcs" Dec 06 16:00:43 crc kubenswrapper[4813]: I1206 16:00:43.634506 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9vgp\" (UniqueName: \"kubernetes.io/projected/e9cd8622-0e78-400c-9de3-213fe41d1c1a-kube-api-access-n9vgp\") pod \"redhat-operators-g8tcs\" (UID: \"e9cd8622-0e78-400c-9de3-213fe41d1c1a\") " pod="openshift-marketplace/redhat-operators-g8tcs" Dec 06 16:00:43 crc kubenswrapper[4813]: I1206 16:00:43.634538 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9cd8622-0e78-400c-9de3-213fe41d1c1a-utilities\") pod \"redhat-operators-g8tcs\" (UID: \"e9cd8622-0e78-400c-9de3-213fe41d1c1a\") " pod="openshift-marketplace/redhat-operators-g8tcs" Dec 06 16:00:43 crc kubenswrapper[4813]: I1206 16:00:43.634965 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9cd8622-0e78-400c-9de3-213fe41d1c1a-utilities\") pod \"redhat-operators-g8tcs\" (UID: \"e9cd8622-0e78-400c-9de3-213fe41d1c1a\") " pod="openshift-marketplace/redhat-operators-g8tcs" Dec 06 16:00:43 crc kubenswrapper[4813]: I1206 16:00:43.635967 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/788502af-b07e-47a4-ad30-3b0181a8104c-utilities" (OuterVolumeSpecName: "utilities") pod "788502af-b07e-47a4-ad30-3b0181a8104c" (UID: "788502af-b07e-47a4-ad30-3b0181a8104c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:00:43 crc kubenswrapper[4813]: I1206 16:00:43.636017 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9cd8622-0e78-400c-9de3-213fe41d1c1a-catalog-content\") pod \"redhat-operators-g8tcs\" (UID: \"e9cd8622-0e78-400c-9de3-213fe41d1c1a\") " pod="openshift-marketplace/redhat-operators-g8tcs" Dec 06 16:00:43 crc kubenswrapper[4813]: I1206 16:00:43.639945 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/788502af-b07e-47a4-ad30-3b0181a8104c-kube-api-access-x66j2" (OuterVolumeSpecName: "kube-api-access-x66j2") pod "788502af-b07e-47a4-ad30-3b0181a8104c" (UID: "788502af-b07e-47a4-ad30-3b0181a8104c"). InnerVolumeSpecName "kube-api-access-x66j2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:00:43 crc kubenswrapper[4813]: I1206 16:00:43.663456 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9vgp\" (UniqueName: \"kubernetes.io/projected/e9cd8622-0e78-400c-9de3-213fe41d1c1a-kube-api-access-n9vgp\") pod \"redhat-operators-g8tcs\" (UID: \"e9cd8622-0e78-400c-9de3-213fe41d1c1a\") " pod="openshift-marketplace/redhat-operators-g8tcs" Dec 06 16:00:43 crc kubenswrapper[4813]: I1206 16:00:43.682540 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8tcs" Dec 06 16:00:43 crc kubenswrapper[4813]: I1206 16:00:43.686389 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/788502af-b07e-47a4-ad30-3b0181a8104c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "788502af-b07e-47a4-ad30-3b0181a8104c" (UID: "788502af-b07e-47a4-ad30-3b0181a8104c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:00:43 crc kubenswrapper[4813]: I1206 16:00:43.735798 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x66j2\" (UniqueName: \"kubernetes.io/projected/788502af-b07e-47a4-ad30-3b0181a8104c-kube-api-access-x66j2\") on node \"crc\" DevicePath \"\"" Dec 06 16:00:43 crc kubenswrapper[4813]: I1206 16:00:43.735829 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/788502af-b07e-47a4-ad30-3b0181a8104c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 16:00:43 crc kubenswrapper[4813]: I1206 16:00:43.735838 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/788502af-b07e-47a4-ad30-3b0181a8104c-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 16:00:43 crc kubenswrapper[4813]: I1206 16:00:43.975645 4813 generic.go:334] "Generic (PLEG): container finished" podID="788502af-b07e-47a4-ad30-3b0181a8104c" containerID="e0c40a65e0b77b681e5c4eababbfe34a203943873bb0c4b3a29fac53c9be75a0" exitCode=0 Dec 06 16:00:43 crc kubenswrapper[4813]: I1206 16:00:43.975708 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ckh76" Dec 06 16:00:43 crc kubenswrapper[4813]: I1206 16:00:43.975713 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ckh76" event={"ID":"788502af-b07e-47a4-ad30-3b0181a8104c","Type":"ContainerDied","Data":"e0c40a65e0b77b681e5c4eababbfe34a203943873bb0c4b3a29fac53c9be75a0"} Dec 06 16:00:43 crc kubenswrapper[4813]: I1206 16:00:43.976728 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ckh76" event={"ID":"788502af-b07e-47a4-ad30-3b0181a8104c","Type":"ContainerDied","Data":"a63ec2127060213fbf39e0e5c5fbb9d34e6d52cd0b809739f0904d24beb1d69f"} Dec 06 16:00:43 crc kubenswrapper[4813]: I1206 16:00:43.976746 4813 scope.go:117] "RemoveContainer" containerID="e0c40a65e0b77b681e5c4eababbfe34a203943873bb0c4b3a29fac53c9be75a0" Dec 06 16:00:43 crc kubenswrapper[4813]: I1206 16:00:43.993694 4813 scope.go:117] "RemoveContainer" containerID="9ac75f149df2eea2bf3748f138d5bc85a05cbb035e8300e01183949956613b67" Dec 06 16:00:44 crc kubenswrapper[4813]: I1206 16:00:44.016336 4813 scope.go:117] "RemoveContainer" containerID="1d208ab8d69a9ba74e68ee73f34291e4a7c37705890ab63473b07eed784332b0" Dec 06 16:00:44 crc kubenswrapper[4813]: I1206 16:00:44.018882 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ckh76"] Dec 06 16:00:44 crc kubenswrapper[4813]: I1206 16:00:44.034434 4813 scope.go:117] "RemoveContainer" containerID="e0c40a65e0b77b681e5c4eababbfe34a203943873bb0c4b3a29fac53c9be75a0" Dec 06 16:00:44 crc kubenswrapper[4813]: I1206 16:00:44.037528 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ckh76"] Dec 06 16:00:44 crc kubenswrapper[4813]: E1206 16:00:44.037944 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0c40a65e0b77b681e5c4eababbfe34a203943873bb0c4b3a29fac53c9be75a0\": container with ID starting with e0c40a65e0b77b681e5c4eababbfe34a203943873bb0c4b3a29fac53c9be75a0 not found: ID does not exist" containerID="e0c40a65e0b77b681e5c4eababbfe34a203943873bb0c4b3a29fac53c9be75a0" Dec 06 16:00:44 crc kubenswrapper[4813]: I1206 16:00:44.037980 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0c40a65e0b77b681e5c4eababbfe34a203943873bb0c4b3a29fac53c9be75a0"} err="failed to get container status \"e0c40a65e0b77b681e5c4eababbfe34a203943873bb0c4b3a29fac53c9be75a0\": rpc error: code = NotFound desc = could not find container \"e0c40a65e0b77b681e5c4eababbfe34a203943873bb0c4b3a29fac53c9be75a0\": container with ID starting with e0c40a65e0b77b681e5c4eababbfe34a203943873bb0c4b3a29fac53c9be75a0 not found: ID does not exist" Dec 06 16:00:44 crc kubenswrapper[4813]: I1206 16:00:44.038001 4813 scope.go:117] "RemoveContainer" containerID="9ac75f149df2eea2bf3748f138d5bc85a05cbb035e8300e01183949956613b67" Dec 06 16:00:44 crc kubenswrapper[4813]: E1206 16:00:44.038948 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ac75f149df2eea2bf3748f138d5bc85a05cbb035e8300e01183949956613b67\": container with ID starting with 9ac75f149df2eea2bf3748f138d5bc85a05cbb035e8300e01183949956613b67 not found: ID does not exist" containerID="9ac75f149df2eea2bf3748f138d5bc85a05cbb035e8300e01183949956613b67" Dec 06 16:00:44 crc kubenswrapper[4813]: I1206 16:00:44.038969 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ac75f149df2eea2bf3748f138d5bc85a05cbb035e8300e01183949956613b67"} err="failed to get container status \"9ac75f149df2eea2bf3748f138d5bc85a05cbb035e8300e01183949956613b67\": rpc error: code = NotFound desc = could not find container \"9ac75f149df2eea2bf3748f138d5bc85a05cbb035e8300e01183949956613b67\": container with ID starting with 9ac75f149df2eea2bf3748f138d5bc85a05cbb035e8300e01183949956613b67 not found: ID does not exist" Dec 06 16:00:44 crc kubenswrapper[4813]: I1206 16:00:44.038984 4813 scope.go:117] "RemoveContainer" containerID="1d208ab8d69a9ba74e68ee73f34291e4a7c37705890ab63473b07eed784332b0" Dec 06 16:00:44 crc kubenswrapper[4813]: E1206 16:00:44.039281 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d208ab8d69a9ba74e68ee73f34291e4a7c37705890ab63473b07eed784332b0\": container with ID starting with 1d208ab8d69a9ba74e68ee73f34291e4a7c37705890ab63473b07eed784332b0 not found: ID does not exist" containerID="1d208ab8d69a9ba74e68ee73f34291e4a7c37705890ab63473b07eed784332b0" Dec 06 16:00:44 crc kubenswrapper[4813]: I1206 16:00:44.039320 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d208ab8d69a9ba74e68ee73f34291e4a7c37705890ab63473b07eed784332b0"} err="failed to get container status \"1d208ab8d69a9ba74e68ee73f34291e4a7c37705890ab63473b07eed784332b0\": rpc error: code = NotFound desc = could not find container \"1d208ab8d69a9ba74e68ee73f34291e4a7c37705890ab63473b07eed784332b0\": container with ID starting with 1d208ab8d69a9ba74e68ee73f34291e4a7c37705890ab63473b07eed784332b0 not found: ID does not exist" Dec 06 16:00:44 crc kubenswrapper[4813]: I1206 16:00:44.211647 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g8tcs"] Dec 06 16:00:44 crc kubenswrapper[4813]: I1206 16:00:44.495238 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="788502af-b07e-47a4-ad30-3b0181a8104c" path="/var/lib/kubelet/pods/788502af-b07e-47a4-ad30-3b0181a8104c/volumes" Dec 06 16:00:44 crc kubenswrapper[4813]: I1206 16:00:44.984418 4813 generic.go:334] "Generic (PLEG): container finished" podID="e9cd8622-0e78-400c-9de3-213fe41d1c1a" containerID="fb08cadb487a1b92cd6fee118777a85510e021c071df300f847f8b323495e72b" exitCode=0 Dec 06 16:00:44 crc kubenswrapper[4813]: I1206 16:00:44.984480 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8tcs" event={"ID":"e9cd8622-0e78-400c-9de3-213fe41d1c1a","Type":"ContainerDied","Data":"fb08cadb487a1b92cd6fee118777a85510e021c071df300f847f8b323495e72b"} Dec 06 16:00:44 crc kubenswrapper[4813]: I1206 16:00:44.984513 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8tcs" event={"ID":"e9cd8622-0e78-400c-9de3-213fe41d1c1a","Type":"ContainerStarted","Data":"4d85bf2e9e73accdca27189247fd184fe4fbd461b46a0b6fe18d7c8c043929e3"} Dec 06 16:00:45 crc kubenswrapper[4813]: I1206 16:00:45.992840 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8tcs" event={"ID":"e9cd8622-0e78-400c-9de3-213fe41d1c1a","Type":"ContainerStarted","Data":"4c2c8620754fe18f4e2b2a097f9aabf2003489a56a3a8ae3aea6ca46c4f3f211"} Dec 06 16:00:47 crc kubenswrapper[4813]: I1206 16:00:47.002134 4813 generic.go:334] "Generic (PLEG): container finished" podID="e9cd8622-0e78-400c-9de3-213fe41d1c1a" containerID="4c2c8620754fe18f4e2b2a097f9aabf2003489a56a3a8ae3aea6ca46c4f3f211" exitCode=0 Dec 06 16:00:47 crc kubenswrapper[4813]: I1206 16:00:47.002197 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8tcs" event={"ID":"e9cd8622-0e78-400c-9de3-213fe41d1c1a","Type":"ContainerDied","Data":"4c2c8620754fe18f4e2b2a097f9aabf2003489a56a3a8ae3aea6ca46c4f3f211"} Dec 06 16:00:49 crc kubenswrapper[4813]: I1206 16:00:49.017782 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8tcs" event={"ID":"e9cd8622-0e78-400c-9de3-213fe41d1c1a","Type":"ContainerStarted","Data":"a51bf28dba637d2e63af3ea9134313fcfecad4d56995810e2c33cddbd6b62283"} Dec 06 16:00:49 crc kubenswrapper[4813]: I1206 16:00:49.550284 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p2k89"] Dec 06 16:00:49 crc kubenswrapper[4813]: E1206 16:00:49.550611 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="788502af-b07e-47a4-ad30-3b0181a8104c" containerName="registry-server" Dec 06 16:00:49 crc kubenswrapper[4813]: I1206 16:00:49.550627 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="788502af-b07e-47a4-ad30-3b0181a8104c" containerName="registry-server" Dec 06 16:00:49 crc kubenswrapper[4813]: E1206 16:00:49.550655 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="788502af-b07e-47a4-ad30-3b0181a8104c" containerName="extract-content" Dec 06 16:00:49 crc kubenswrapper[4813]: I1206 16:00:49.550663 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="788502af-b07e-47a4-ad30-3b0181a8104c" containerName="extract-content" Dec 06 16:00:49 crc kubenswrapper[4813]: E1206 16:00:49.550688 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="788502af-b07e-47a4-ad30-3b0181a8104c" containerName="extract-utilities" Dec 06 16:00:49 crc kubenswrapper[4813]: I1206 16:00:49.550697 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="788502af-b07e-47a4-ad30-3b0181a8104c" containerName="extract-utilities" Dec 06 16:00:49 crc kubenswrapper[4813]: I1206 16:00:49.550864 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="788502af-b07e-47a4-ad30-3b0181a8104c" containerName="registry-server" Dec 06 16:00:49 crc kubenswrapper[4813]: I1206 16:00:49.551805 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p2k89" Dec 06 16:00:49 crc kubenswrapper[4813]: I1206 16:00:49.566842 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p2k89"] Dec 06 16:00:49 crc kubenswrapper[4813]: I1206 16:00:49.626041 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7c84964-2748-489e-9397-750792497b19-catalog-content\") pod \"certified-operators-p2k89\" (UID: \"d7c84964-2748-489e-9397-750792497b19\") " pod="openshift-marketplace/certified-operators-p2k89" Dec 06 16:00:49 crc kubenswrapper[4813]: I1206 16:00:49.626116 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnhb6\" (UniqueName: \"kubernetes.io/projected/d7c84964-2748-489e-9397-750792497b19-kube-api-access-gnhb6\") pod \"certified-operators-p2k89\" (UID: \"d7c84964-2748-489e-9397-750792497b19\") " pod="openshift-marketplace/certified-operators-p2k89" Dec 06 16:00:49 crc kubenswrapper[4813]: I1206 16:00:49.626173 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7c84964-2748-489e-9397-750792497b19-utilities\") pod \"certified-operators-p2k89\" (UID: \"d7c84964-2748-489e-9397-750792497b19\") " pod="openshift-marketplace/certified-operators-p2k89" Dec 06 16:00:49 crc kubenswrapper[4813]: I1206 16:00:49.727701 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnhb6\" (UniqueName: \"kubernetes.io/projected/d7c84964-2748-489e-9397-750792497b19-kube-api-access-gnhb6\") pod \"certified-operators-p2k89\" (UID: \"d7c84964-2748-489e-9397-750792497b19\") " pod="openshift-marketplace/certified-operators-p2k89" Dec 06 16:00:49 crc kubenswrapper[4813]: I1206 16:00:49.727804 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7c84964-2748-489e-9397-750792497b19-utilities\") pod \"certified-operators-p2k89\" (UID: \"d7c84964-2748-489e-9397-750792497b19\") " pod="openshift-marketplace/certified-operators-p2k89" Dec 06 16:00:49 crc kubenswrapper[4813]: I1206 16:00:49.727884 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7c84964-2748-489e-9397-750792497b19-catalog-content\") pod \"certified-operators-p2k89\" (UID: \"d7c84964-2748-489e-9397-750792497b19\") " pod="openshift-marketplace/certified-operators-p2k89" Dec 06 16:00:49 crc kubenswrapper[4813]: I1206 16:00:49.728491 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7c84964-2748-489e-9397-750792497b19-catalog-content\") pod \"certified-operators-p2k89\" (UID: \"d7c84964-2748-489e-9397-750792497b19\") " pod="openshift-marketplace/certified-operators-p2k89" Dec 06 16:00:49 crc kubenswrapper[4813]: I1206 16:00:49.729167 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7c84964-2748-489e-9397-750792497b19-utilities\") pod \"certified-operators-p2k89\" (UID: \"d7c84964-2748-489e-9397-750792497b19\") " pod="openshift-marketplace/certified-operators-p2k89" Dec 06 16:00:49 crc kubenswrapper[4813]: I1206 16:00:49.748020 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnhb6\" (UniqueName: \"kubernetes.io/projected/d7c84964-2748-489e-9397-750792497b19-kube-api-access-gnhb6\") pod \"certified-operators-p2k89\" (UID: \"d7c84964-2748-489e-9397-750792497b19\") " pod="openshift-marketplace/certified-operators-p2k89" Dec 06 16:00:49 crc kubenswrapper[4813]: I1206 16:00:49.883189 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p2k89" Dec 06 16:00:50 crc kubenswrapper[4813]: I1206 16:00:50.127070 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-g8tcs" podStartSLOduration=4.432554472 podStartE2EDuration="7.127050161s" podCreationTimestamp="2025-12-06 16:00:43 +0000 UTC" firstStartedPulling="2025-12-06 16:00:44.985836238 +0000 UTC m=+884.876715814" lastFinishedPulling="2025-12-06 16:00:47.680331917 +0000 UTC m=+887.571211503" observedRunningTime="2025-12-06 16:00:50.093828551 +0000 UTC m=+889.984708127" watchObservedRunningTime="2025-12-06 16:00:50.127050161 +0000 UTC m=+890.017929737" Dec 06 16:00:50 crc kubenswrapper[4813]: W1206 16:00:50.406324 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd7c84964_2748_489e_9397_750792497b19.slice/crio-a091e7bd77170f359e5f4e037d779f45b3d0bec2b598d17d476f82b927f3ecb1 WatchSource:0}: Error finding container a091e7bd77170f359e5f4e037d779f45b3d0bec2b598d17d476f82b927f3ecb1: Status 404 returned error can't find the container with id a091e7bd77170f359e5f4e037d779f45b3d0bec2b598d17d476f82b927f3ecb1 Dec 06 16:00:50 crc kubenswrapper[4813]: I1206 16:00:50.407184 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p2k89"] Dec 06 16:00:51 crc kubenswrapper[4813]: I1206 16:00:51.058039 4813 generic.go:334] "Generic (PLEG): container finished" podID="d7c84964-2748-489e-9397-750792497b19" containerID="cad225f5f92c252eb9a79a6b20de04be0ee9c7ba0dbb523d072c16933bc22b23" exitCode=0 Dec 06 16:00:51 crc kubenswrapper[4813]: I1206 16:00:51.058095 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p2k89" event={"ID":"d7c84964-2748-489e-9397-750792497b19","Type":"ContainerDied","Data":"cad225f5f92c252eb9a79a6b20de04be0ee9c7ba0dbb523d072c16933bc22b23"} Dec 06 16:00:51 crc kubenswrapper[4813]: I1206 16:00:51.058137 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p2k89" event={"ID":"d7c84964-2748-489e-9397-750792497b19","Type":"ContainerStarted","Data":"a091e7bd77170f359e5f4e037d779f45b3d0bec2b598d17d476f82b927f3ecb1"} Dec 06 16:00:52 crc kubenswrapper[4813]: I1206 16:00:52.067617 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p2k89" event={"ID":"d7c84964-2748-489e-9397-750792497b19","Type":"ContainerStarted","Data":"38ed1e7a29dae2d294373293ba666fce3ea4a33c96e8e130f8a1cb866504b1c6"} Dec 06 16:00:53 crc kubenswrapper[4813]: I1206 16:00:53.091988 4813 generic.go:334] "Generic (PLEG): container finished" podID="d7c84964-2748-489e-9397-750792497b19" containerID="38ed1e7a29dae2d294373293ba666fce3ea4a33c96e8e130f8a1cb866504b1c6" exitCode=0 Dec 06 16:00:53 crc kubenswrapper[4813]: I1206 16:00:53.092100 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p2k89" event={"ID":"d7c84964-2748-489e-9397-750792497b19","Type":"ContainerDied","Data":"38ed1e7a29dae2d294373293ba666fce3ea4a33c96e8e130f8a1cb866504b1c6"} Dec 06 16:00:53 crc kubenswrapper[4813]: I1206 16:00:53.683670 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-g8tcs" Dec 06 16:00:53 crc kubenswrapper[4813]: I1206 16:00:53.683718 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-g8tcs" Dec 06 16:00:54 crc kubenswrapper[4813]: I1206 16:00:54.101525 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p2k89" event={"ID":"d7c84964-2748-489e-9397-750792497b19","Type":"ContainerStarted","Data":"8010dc56575a8c0de633f5111b9c41543e7fa4be4d19dcdff3279ae3b42501a2"} Dec 06 16:00:54 crc kubenswrapper[4813]: I1206 16:00:54.121351 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p2k89" podStartSLOduration=2.5994402450000003 podStartE2EDuration="5.121331599s" podCreationTimestamp="2025-12-06 16:00:49 +0000 UTC" firstStartedPulling="2025-12-06 16:00:51.059794822 +0000 UTC m=+890.950674398" lastFinishedPulling="2025-12-06 16:00:53.581686136 +0000 UTC m=+893.472565752" observedRunningTime="2025-12-06 16:00:54.117969724 +0000 UTC m=+894.008849300" watchObservedRunningTime="2025-12-06 16:00:54.121331599 +0000 UTC m=+894.012211175" Dec 06 16:00:54 crc kubenswrapper[4813]: I1206 16:00:54.756911 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-g8tcs" podUID="e9cd8622-0e78-400c-9de3-213fe41d1c1a" containerName="registry-server" probeResult="failure" output=< Dec 06 16:00:54 crc kubenswrapper[4813]: timeout: failed to connect service ":50051" within 1s Dec 06 16:00:54 crc kubenswrapper[4813]: > Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.585410 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-nng9s"] Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.586840 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-nng9s" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.588832 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-s94k9" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.588842 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.588843 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.589504 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.591365 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-nng9s"] Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.654228 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-whjs8"] Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.655707 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-whjs8" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.661273 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.672840 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-whjs8"] Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.736747 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e0d82a4-b89e-4c97-9038-e1554d04963b-config\") pod \"dnsmasq-dns-78dd6ddcc-whjs8\" (UID: \"7e0d82a4-b89e-4c97-9038-e1554d04963b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-whjs8" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.736819 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e0d82a4-b89e-4c97-9038-e1554d04963b-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-whjs8\" (UID: \"7e0d82a4-b89e-4c97-9038-e1554d04963b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-whjs8" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.736858 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktsw7\" (UniqueName: \"kubernetes.io/projected/ee786ef0-6a58-4266-ad9e-93a97b9d4cbf-kube-api-access-ktsw7\") pod \"dnsmasq-dns-675f4bcbfc-nng9s\" (UID: \"ee786ef0-6a58-4266-ad9e-93a97b9d4cbf\") " pod="openstack/dnsmasq-dns-675f4bcbfc-nng9s" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.736881 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee786ef0-6a58-4266-ad9e-93a97b9d4cbf-config\") pod \"dnsmasq-dns-675f4bcbfc-nng9s\" (UID: \"ee786ef0-6a58-4266-ad9e-93a97b9d4cbf\") " pod="openstack/dnsmasq-dns-675f4bcbfc-nng9s" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.736913 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb49c\" (UniqueName: \"kubernetes.io/projected/7e0d82a4-b89e-4c97-9038-e1554d04963b-kube-api-access-bb49c\") pod \"dnsmasq-dns-78dd6ddcc-whjs8\" (UID: \"7e0d82a4-b89e-4c97-9038-e1554d04963b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-whjs8" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.747982 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4lwvb"] Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.749353 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4lwvb" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.763836 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4lwvb"] Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.838205 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb49c\" (UniqueName: \"kubernetes.io/projected/7e0d82a4-b89e-4c97-9038-e1554d04963b-kube-api-access-bb49c\") pod \"dnsmasq-dns-78dd6ddcc-whjs8\" (UID: \"7e0d82a4-b89e-4c97-9038-e1554d04963b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-whjs8" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.838270 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f91a370-ae93-4992-95f9-ebb72180adec-utilities\") pod \"redhat-marketplace-4lwvb\" (UID: \"1f91a370-ae93-4992-95f9-ebb72180adec\") " pod="openshift-marketplace/redhat-marketplace-4lwvb" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.838298 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e0d82a4-b89e-4c97-9038-e1554d04963b-config\") pod \"dnsmasq-dns-78dd6ddcc-whjs8\" (UID: \"7e0d82a4-b89e-4c97-9038-e1554d04963b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-whjs8" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.838329 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f91a370-ae93-4992-95f9-ebb72180adec-catalog-content\") pod \"redhat-marketplace-4lwvb\" (UID: \"1f91a370-ae93-4992-95f9-ebb72180adec\") " pod="openshift-marketplace/redhat-marketplace-4lwvb" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.838360 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e0d82a4-b89e-4c97-9038-e1554d04963b-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-whjs8\" (UID: \"7e0d82a4-b89e-4c97-9038-e1554d04963b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-whjs8" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.838390 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfsj2\" (UniqueName: \"kubernetes.io/projected/1f91a370-ae93-4992-95f9-ebb72180adec-kube-api-access-wfsj2\") pod \"redhat-marketplace-4lwvb\" (UID: \"1f91a370-ae93-4992-95f9-ebb72180adec\") " pod="openshift-marketplace/redhat-marketplace-4lwvb" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.838412 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktsw7\" (UniqueName: \"kubernetes.io/projected/ee786ef0-6a58-4266-ad9e-93a97b9d4cbf-kube-api-access-ktsw7\") pod \"dnsmasq-dns-675f4bcbfc-nng9s\" (UID: \"ee786ef0-6a58-4266-ad9e-93a97b9d4cbf\") " pod="openstack/dnsmasq-dns-675f4bcbfc-nng9s" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.838431 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee786ef0-6a58-4266-ad9e-93a97b9d4cbf-config\") pod \"dnsmasq-dns-675f4bcbfc-nng9s\" (UID: \"ee786ef0-6a58-4266-ad9e-93a97b9d4cbf\") " pod="openstack/dnsmasq-dns-675f4bcbfc-nng9s" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.839214 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee786ef0-6a58-4266-ad9e-93a97b9d4cbf-config\") pod \"dnsmasq-dns-675f4bcbfc-nng9s\" (UID: \"ee786ef0-6a58-4266-ad9e-93a97b9d4cbf\") " pod="openstack/dnsmasq-dns-675f4bcbfc-nng9s" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.839717 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e0d82a4-b89e-4c97-9038-e1554d04963b-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-whjs8\" (UID: \"7e0d82a4-b89e-4c97-9038-e1554d04963b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-whjs8" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.840333 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e0d82a4-b89e-4c97-9038-e1554d04963b-config\") pod \"dnsmasq-dns-78dd6ddcc-whjs8\" (UID: \"7e0d82a4-b89e-4c97-9038-e1554d04963b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-whjs8" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.864684 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb49c\" (UniqueName: \"kubernetes.io/projected/7e0d82a4-b89e-4c97-9038-e1554d04963b-kube-api-access-bb49c\") pod \"dnsmasq-dns-78dd6ddcc-whjs8\" (UID: \"7e0d82a4-b89e-4c97-9038-e1554d04963b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-whjs8" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.871833 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktsw7\" (UniqueName: \"kubernetes.io/projected/ee786ef0-6a58-4266-ad9e-93a97b9d4cbf-kube-api-access-ktsw7\") pod \"dnsmasq-dns-675f4bcbfc-nng9s\" (UID: \"ee786ef0-6a58-4266-ad9e-93a97b9d4cbf\") " pod="openstack/dnsmasq-dns-675f4bcbfc-nng9s" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.904166 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-nng9s" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.950459 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f91a370-ae93-4992-95f9-ebb72180adec-catalog-content\") pod \"redhat-marketplace-4lwvb\" (UID: \"1f91a370-ae93-4992-95f9-ebb72180adec\") " pod="openshift-marketplace/redhat-marketplace-4lwvb" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.950548 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfsj2\" (UniqueName: \"kubernetes.io/projected/1f91a370-ae93-4992-95f9-ebb72180adec-kube-api-access-wfsj2\") pod \"redhat-marketplace-4lwvb\" (UID: \"1f91a370-ae93-4992-95f9-ebb72180adec\") " pod="openshift-marketplace/redhat-marketplace-4lwvb" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.950617 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f91a370-ae93-4992-95f9-ebb72180adec-utilities\") pod \"redhat-marketplace-4lwvb\" (UID: \"1f91a370-ae93-4992-95f9-ebb72180adec\") " pod="openshift-marketplace/redhat-marketplace-4lwvb" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.951094 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f91a370-ae93-4992-95f9-ebb72180adec-utilities\") pod \"redhat-marketplace-4lwvb\" (UID: \"1f91a370-ae93-4992-95f9-ebb72180adec\") " pod="openshift-marketplace/redhat-marketplace-4lwvb" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.951169 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f91a370-ae93-4992-95f9-ebb72180adec-catalog-content\") pod \"redhat-marketplace-4lwvb\" (UID: \"1f91a370-ae93-4992-95f9-ebb72180adec\") " pod="openshift-marketplace/redhat-marketplace-4lwvb" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.982574 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-whjs8" Dec 06 16:00:55 crc kubenswrapper[4813]: I1206 16:00:55.989125 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfsj2\" (UniqueName: \"kubernetes.io/projected/1f91a370-ae93-4992-95f9-ebb72180adec-kube-api-access-wfsj2\") pod \"redhat-marketplace-4lwvb\" (UID: \"1f91a370-ae93-4992-95f9-ebb72180adec\") " pod="openshift-marketplace/redhat-marketplace-4lwvb" Dec 06 16:00:56 crc kubenswrapper[4813]: I1206 16:00:56.062242 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4lwvb" Dec 06 16:00:56 crc kubenswrapper[4813]: I1206 16:00:56.540710 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-nng9s"] Dec 06 16:00:56 crc kubenswrapper[4813]: I1206 16:00:56.555797 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4lwvb"] Dec 06 16:00:56 crc kubenswrapper[4813]: I1206 16:00:56.660193 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-whjs8"] Dec 06 16:00:56 crc kubenswrapper[4813]: W1206 16:00:56.671257 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e0d82a4_b89e_4c97_9038_e1554d04963b.slice/crio-a1d1146607cb64ac90ee024db2a392925639d246fe6142bd69f43ddcffe8dee6 WatchSource:0}: Error finding container a1d1146607cb64ac90ee024db2a392925639d246fe6142bd69f43ddcffe8dee6: Status 404 returned error can't find the container with id a1d1146607cb64ac90ee024db2a392925639d246fe6142bd69f43ddcffe8dee6 Dec 06 16:00:57 crc kubenswrapper[4813]: I1206 16:00:57.138472 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-nng9s" event={"ID":"ee786ef0-6a58-4266-ad9e-93a97b9d4cbf","Type":"ContainerStarted","Data":"160bcf4587ef9b3e65ef92be86c9643f98cbe7adaeb14eeb91479c3979e79a82"} Dec 06 16:00:57 crc kubenswrapper[4813]: I1206 16:00:57.141230 4813 generic.go:334] "Generic (PLEG): container finished" podID="1f91a370-ae93-4992-95f9-ebb72180adec" containerID="f2897eba0d29b862597ab3dac690c7afc0393fd65c5f8ac802687c8113c7c543" exitCode=0 Dec 06 16:00:57 crc kubenswrapper[4813]: I1206 16:00:57.141981 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lwvb" event={"ID":"1f91a370-ae93-4992-95f9-ebb72180adec","Type":"ContainerDied","Data":"f2897eba0d29b862597ab3dac690c7afc0393fd65c5f8ac802687c8113c7c543"} Dec 06 16:00:57 crc kubenswrapper[4813]: I1206 16:00:57.142013 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lwvb" event={"ID":"1f91a370-ae93-4992-95f9-ebb72180adec","Type":"ContainerStarted","Data":"27895c62025fb9cb7e34371bb4209e4a4b34e5307c101887e00d46a6ec0ea7d8"} Dec 06 16:00:57 crc kubenswrapper[4813]: I1206 16:00:57.149706 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-whjs8" event={"ID":"7e0d82a4-b89e-4c97-9038-e1554d04963b","Type":"ContainerStarted","Data":"a1d1146607cb64ac90ee024db2a392925639d246fe6142bd69f43ddcffe8dee6"} Dec 06 16:00:58 crc kubenswrapper[4813]: I1206 16:00:58.701877 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-nng9s"] Dec 06 16:00:58 crc kubenswrapper[4813]: I1206 16:00:58.750161 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-2fqgs"] Dec 06 16:00:58 crc kubenswrapper[4813]: I1206 16:00:58.751578 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-2fqgs" Dec 06 16:00:58 crc kubenswrapper[4813]: I1206 16:00:58.768333 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-2fqgs"] Dec 06 16:00:58 crc kubenswrapper[4813]: I1206 16:00:58.898769 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggxdn\" (UniqueName: \"kubernetes.io/projected/00d42d29-5c9b-4585-be1b-51077e528837-kube-api-access-ggxdn\") pod \"dnsmasq-dns-5ccc8479f9-2fqgs\" (UID: \"00d42d29-5c9b-4585-be1b-51077e528837\") " pod="openstack/dnsmasq-dns-5ccc8479f9-2fqgs" Dec 06 16:00:58 crc kubenswrapper[4813]: I1206 16:00:58.899151 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00d42d29-5c9b-4585-be1b-51077e528837-config\") pod \"dnsmasq-dns-5ccc8479f9-2fqgs\" (UID: \"00d42d29-5c9b-4585-be1b-51077e528837\") " pod="openstack/dnsmasq-dns-5ccc8479f9-2fqgs" Dec 06 16:00:58 crc kubenswrapper[4813]: I1206 16:00:58.899178 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00d42d29-5c9b-4585-be1b-51077e528837-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-2fqgs\" (UID: \"00d42d29-5c9b-4585-be1b-51077e528837\") " pod="openstack/dnsmasq-dns-5ccc8479f9-2fqgs" Dec 06 16:00:59 crc kubenswrapper[4813]: I1206 16:00:59.000725 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggxdn\" (UniqueName: \"kubernetes.io/projected/00d42d29-5c9b-4585-be1b-51077e528837-kube-api-access-ggxdn\") pod \"dnsmasq-dns-5ccc8479f9-2fqgs\" (UID: \"00d42d29-5c9b-4585-be1b-51077e528837\") " pod="openstack/dnsmasq-dns-5ccc8479f9-2fqgs" Dec 06 16:00:59 crc kubenswrapper[4813]: I1206 16:00:59.000836 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00d42d29-5c9b-4585-be1b-51077e528837-config\") pod \"dnsmasq-dns-5ccc8479f9-2fqgs\" (UID: \"00d42d29-5c9b-4585-be1b-51077e528837\") " pod="openstack/dnsmasq-dns-5ccc8479f9-2fqgs" Dec 06 16:00:59 crc kubenswrapper[4813]: I1206 16:00:59.000865 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00d42d29-5c9b-4585-be1b-51077e528837-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-2fqgs\" (UID: \"00d42d29-5c9b-4585-be1b-51077e528837\") " pod="openstack/dnsmasq-dns-5ccc8479f9-2fqgs" Dec 06 16:00:59 crc kubenswrapper[4813]: I1206 16:00:59.002092 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00d42d29-5c9b-4585-be1b-51077e528837-config\") pod \"dnsmasq-dns-5ccc8479f9-2fqgs\" (UID: \"00d42d29-5c9b-4585-be1b-51077e528837\") " pod="openstack/dnsmasq-dns-5ccc8479f9-2fqgs" Dec 06 16:00:59 crc kubenswrapper[4813]: I1206 16:00:59.002122 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00d42d29-5c9b-4585-be1b-51077e528837-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-2fqgs\" (UID: \"00d42d29-5c9b-4585-be1b-51077e528837\") " pod="openstack/dnsmasq-dns-5ccc8479f9-2fqgs" Dec 06 16:00:59 crc kubenswrapper[4813]: I1206 16:00:59.021989 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggxdn\" (UniqueName: \"kubernetes.io/projected/00d42d29-5c9b-4585-be1b-51077e528837-kube-api-access-ggxdn\") pod \"dnsmasq-dns-5ccc8479f9-2fqgs\" (UID: \"00d42d29-5c9b-4585-be1b-51077e528837\") " pod="openstack/dnsmasq-dns-5ccc8479f9-2fqgs" Dec 06 16:00:59 crc kubenswrapper[4813]: I1206 16:00:59.089654 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-2fqgs" Dec 06 16:00:59 crc kubenswrapper[4813]: I1206 16:00:59.130060 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-whjs8"] Dec 06 16:00:59 crc kubenswrapper[4813]: I1206 16:00:59.172451 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lwvb" event={"ID":"1f91a370-ae93-4992-95f9-ebb72180adec","Type":"ContainerStarted","Data":"15569de231f50cea091f44a21f523150826f61f883ffb6bacb50bb429212b636"} Dec 06 16:00:59 crc kubenswrapper[4813]: I1206 16:00:59.176320 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-6tgdq"] Dec 06 16:00:59 crc kubenswrapper[4813]: I1206 16:00:59.177497 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-6tgdq" Dec 06 16:00:59 crc kubenswrapper[4813]: I1206 16:00:59.199002 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-6tgdq"] Dec 06 16:00:59 crc kubenswrapper[4813]: I1206 16:00:59.312621 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb685be0-8427-43c2-ac4e-262c73714b5f-config\") pod \"dnsmasq-dns-57d769cc4f-6tgdq\" (UID: \"fb685be0-8427-43c2-ac4e-262c73714b5f\") " pod="openstack/dnsmasq-dns-57d769cc4f-6tgdq" Dec 06 16:00:59 crc kubenswrapper[4813]: I1206 16:00:59.312701 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5b5x\" (UniqueName: \"kubernetes.io/projected/fb685be0-8427-43c2-ac4e-262c73714b5f-kube-api-access-d5b5x\") pod \"dnsmasq-dns-57d769cc4f-6tgdq\" (UID: \"fb685be0-8427-43c2-ac4e-262c73714b5f\") " pod="openstack/dnsmasq-dns-57d769cc4f-6tgdq" Dec 06 16:00:59 crc kubenswrapper[4813]: I1206 16:00:59.312773 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb685be0-8427-43c2-ac4e-262c73714b5f-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-6tgdq\" (UID: \"fb685be0-8427-43c2-ac4e-262c73714b5f\") " pod="openstack/dnsmasq-dns-57d769cc4f-6tgdq" Dec 06 16:00:59 crc kubenswrapper[4813]: I1206 16:00:59.415151 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb685be0-8427-43c2-ac4e-262c73714b5f-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-6tgdq\" (UID: \"fb685be0-8427-43c2-ac4e-262c73714b5f\") " pod="openstack/dnsmasq-dns-57d769cc4f-6tgdq" Dec 06 16:00:59 crc kubenswrapper[4813]: I1206 16:00:59.417280 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb685be0-8427-43c2-ac4e-262c73714b5f-config\") pod \"dnsmasq-dns-57d769cc4f-6tgdq\" (UID: \"fb685be0-8427-43c2-ac4e-262c73714b5f\") " pod="openstack/dnsmasq-dns-57d769cc4f-6tgdq" Dec 06 16:00:59 crc kubenswrapper[4813]: I1206 16:00:59.417345 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5b5x\" (UniqueName: \"kubernetes.io/projected/fb685be0-8427-43c2-ac4e-262c73714b5f-kube-api-access-d5b5x\") pod \"dnsmasq-dns-57d769cc4f-6tgdq\" (UID: \"fb685be0-8427-43c2-ac4e-262c73714b5f\") " pod="openstack/dnsmasq-dns-57d769cc4f-6tgdq" Dec 06 16:00:59 crc kubenswrapper[4813]: I1206 16:00:59.418056 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb685be0-8427-43c2-ac4e-262c73714b5f-config\") pod \"dnsmasq-dns-57d769cc4f-6tgdq\" (UID: \"fb685be0-8427-43c2-ac4e-262c73714b5f\") " pod="openstack/dnsmasq-dns-57d769cc4f-6tgdq" Dec 06 16:00:59 crc kubenswrapper[4813]: I1206 16:00:59.421481 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb685be0-8427-43c2-ac4e-262c73714b5f-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-6tgdq\" (UID: \"fb685be0-8427-43c2-ac4e-262c73714b5f\") " pod="openstack/dnsmasq-dns-57d769cc4f-6tgdq" Dec 06 16:00:59 crc kubenswrapper[4813]: I1206 16:00:59.442203 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5b5x\" (UniqueName: \"kubernetes.io/projected/fb685be0-8427-43c2-ac4e-262c73714b5f-kube-api-access-d5b5x\") pod \"dnsmasq-dns-57d769cc4f-6tgdq\" (UID: \"fb685be0-8427-43c2-ac4e-262c73714b5f\") " pod="openstack/dnsmasq-dns-57d769cc4f-6tgdq" Dec 06 16:00:59 crc kubenswrapper[4813]: I1206 16:00:59.526992 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-6tgdq" Dec 06 16:00:59 crc kubenswrapper[4813]: I1206 16:00:59.720360 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-2fqgs"] Dec 06 16:00:59 crc kubenswrapper[4813]: W1206 16:00:59.739606 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00d42d29_5c9b_4585_be1b_51077e528837.slice/crio-495e4ccbc38c31136a94ca11755afb0728a50af7fc99d86c23845fc9cf0b6dfd WatchSource:0}: Error finding container 495e4ccbc38c31136a94ca11755afb0728a50af7fc99d86c23845fc9cf0b6dfd: Status 404 returned error can't find the container with id 495e4ccbc38c31136a94ca11755afb0728a50af7fc99d86c23845fc9cf0b6dfd Dec 06 16:00:59 crc kubenswrapper[4813]: I1206 16:00:59.754474 4813 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 16:00:59 crc kubenswrapper[4813]: I1206 16:00:59.883935 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p2k89" Dec 06 16:00:59 crc kubenswrapper[4813]: I1206 16:00:59.884309 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p2k89" Dec 06 16:00:59 crc kubenswrapper[4813]: I1206 16:00:59.945871 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p2k89" Dec 06 16:00:59 crc kubenswrapper[4813]: I1206 16:00:59.987670 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 16:00:59 crc kubenswrapper[4813]: I1206 16:00:59.994082 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.007487 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.008596 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.008779 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.009014 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-69ctp" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.012280 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.012738 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.025551 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.025695 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.061279 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-6tgdq"] Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.129591 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2172dfb5-c0a6-4525-a326-dd9736eb5233-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.129632 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.129656 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2172dfb5-c0a6-4525-a326-dd9736eb5233-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.129680 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2172dfb5-c0a6-4525-a326-dd9736eb5233-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.129701 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2172dfb5-c0a6-4525-a326-dd9736eb5233-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.129759 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2172dfb5-c0a6-4525-a326-dd9736eb5233-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.129817 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2172dfb5-c0a6-4525-a326-dd9736eb5233-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.129845 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2172dfb5-c0a6-4525-a326-dd9736eb5233-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.129860 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2172dfb5-c0a6-4525-a326-dd9736eb5233-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.129887 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rprv\" (UniqueName: \"kubernetes.io/projected/2172dfb5-c0a6-4525-a326-dd9736eb5233-kube-api-access-4rprv\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.129906 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2172dfb5-c0a6-4525-a326-dd9736eb5233-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.191557 4813 generic.go:334] "Generic (PLEG): container finished" podID="1f91a370-ae93-4992-95f9-ebb72180adec" containerID="15569de231f50cea091f44a21f523150826f61f883ffb6bacb50bb429212b636" exitCode=0 Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.191608 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lwvb" event={"ID":"1f91a370-ae93-4992-95f9-ebb72180adec","Type":"ContainerDied","Data":"15569de231f50cea091f44a21f523150826f61f883ffb6bacb50bb429212b636"} Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.193694 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-2fqgs" event={"ID":"00d42d29-5c9b-4585-be1b-51077e528837","Type":"ContainerStarted","Data":"495e4ccbc38c31136a94ca11755afb0728a50af7fc99d86c23845fc9cf0b6dfd"} Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.196571 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-6tgdq" event={"ID":"fb685be0-8427-43c2-ac4e-262c73714b5f","Type":"ContainerStarted","Data":"6016788dc03ab2801e4b984ea374765bae15fd5d2d6c7fdb1b318c0c3b64ac99"} Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.231965 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2172dfb5-c0a6-4525-a326-dd9736eb5233-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.232005 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2172dfb5-c0a6-4525-a326-dd9736eb5233-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.232021 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2172dfb5-c0a6-4525-a326-dd9736eb5233-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.232081 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rprv\" (UniqueName: \"kubernetes.io/projected/2172dfb5-c0a6-4525-a326-dd9736eb5233-kube-api-access-4rprv\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.232104 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2172dfb5-c0a6-4525-a326-dd9736eb5233-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.232151 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2172dfb5-c0a6-4525-a326-dd9736eb5233-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.232202 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.232224 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2172dfb5-c0a6-4525-a326-dd9736eb5233-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.232248 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2172dfb5-c0a6-4525-a326-dd9736eb5233-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.232281 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2172dfb5-c0a6-4525-a326-dd9736eb5233-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.232301 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2172dfb5-c0a6-4525-a326-dd9736eb5233-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.232684 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.233568 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2172dfb5-c0a6-4525-a326-dd9736eb5233-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.233502 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2172dfb5-c0a6-4525-a326-dd9736eb5233-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.234613 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2172dfb5-c0a6-4525-a326-dd9736eb5233-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.234738 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2172dfb5-c0a6-4525-a326-dd9736eb5233-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.235520 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2172dfb5-c0a6-4525-a326-dd9736eb5233-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.239149 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2172dfb5-c0a6-4525-a326-dd9736eb5233-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.239426 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2172dfb5-c0a6-4525-a326-dd9736eb5233-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.239234 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2172dfb5-c0a6-4525-a326-dd9736eb5233-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.256223 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2172dfb5-c0a6-4525-a326-dd9736eb5233-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.258924 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rprv\" (UniqueName: \"kubernetes.io/projected/2172dfb5-c0a6-4525-a326-dd9736eb5233-kube-api-access-4rprv\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.259105 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p2k89" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.265881 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.331614 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.334517 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.339548 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.341384 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.342185 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.343155 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.343443 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.343537 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-nndc8" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.343581 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.343724 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.361563 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.435036 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.435105 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.435126 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.435153 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.435175 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.435212 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzdk5\" (UniqueName: \"kubernetes.io/projected/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-kube-api-access-rzdk5\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.435237 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.435267 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.435289 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-config-data\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.435308 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.435331 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.539910 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.539968 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.539987 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.540016 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.540033 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.540074 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzdk5\" (UniqueName: \"kubernetes.io/projected/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-kube-api-access-rzdk5\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.540100 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.540118 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.540141 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-config-data\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.540162 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.540182 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.540212 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.543462 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.547914 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.550102 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.554874 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.558866 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.562327 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.562569 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.562746 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.565414 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.568803 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzdk5\" (UniqueName: \"kubernetes.io/projected/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-kube-api-access-rzdk5\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.572282 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.573088 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.575526 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-config-data\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.577860 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.584222 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.588064 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.594139 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " pod="openstack/rabbitmq-server-0" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.685803 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-nndc8" Dec 06 16:01:00 crc kubenswrapper[4813]: I1206 16:01:00.686722 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.541336 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p2k89"] Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.716406 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.718416 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.728167 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.728202 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.728237 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-mkpsz" Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.729413 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.730929 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.742050 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.865192 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/75d94fff-d912-4493-b9b9-c458138a7ccf-config-data-default\") pod \"openstack-galera-0\" (UID: \"75d94fff-d912-4493-b9b9-c458138a7ccf\") " pod="openstack/openstack-galera-0" Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.865248 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t4dw\" (UniqueName: \"kubernetes.io/projected/75d94fff-d912-4493-b9b9-c458138a7ccf-kube-api-access-9t4dw\") pod \"openstack-galera-0\" (UID: \"75d94fff-d912-4493-b9b9-c458138a7ccf\") " pod="openstack/openstack-galera-0" Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.865304 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/75d94fff-d912-4493-b9b9-c458138a7ccf-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"75d94fff-d912-4493-b9b9-c458138a7ccf\") " pod="openstack/openstack-galera-0" Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.865328 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75d94fff-d912-4493-b9b9-c458138a7ccf-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"75d94fff-d912-4493-b9b9-c458138a7ccf\") " pod="openstack/openstack-galera-0" Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.865342 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75d94fff-d912-4493-b9b9-c458138a7ccf-operator-scripts\") pod \"openstack-galera-0\" (UID: \"75d94fff-d912-4493-b9b9-c458138a7ccf\") " pod="openstack/openstack-galera-0" Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.865363 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"75d94fff-d912-4493-b9b9-c458138a7ccf\") " pod="openstack/openstack-galera-0" Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.865403 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/75d94fff-d912-4493-b9b9-c458138a7ccf-config-data-generated\") pod \"openstack-galera-0\" (UID: \"75d94fff-d912-4493-b9b9-c458138a7ccf\") " pod="openstack/openstack-galera-0" Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.865429 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/75d94fff-d912-4493-b9b9-c458138a7ccf-kolla-config\") pod \"openstack-galera-0\" (UID: \"75d94fff-d912-4493-b9b9-c458138a7ccf\") " pod="openstack/openstack-galera-0" Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.968216 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/75d94fff-d912-4493-b9b9-c458138a7ccf-kolla-config\") pod \"openstack-galera-0\" (UID: \"75d94fff-d912-4493-b9b9-c458138a7ccf\") " pod="openstack/openstack-galera-0" Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.968321 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/75d94fff-d912-4493-b9b9-c458138a7ccf-config-data-default\") pod \"openstack-galera-0\" (UID: \"75d94fff-d912-4493-b9b9-c458138a7ccf\") " pod="openstack/openstack-galera-0" Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.968766 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t4dw\" (UniqueName: \"kubernetes.io/projected/75d94fff-d912-4493-b9b9-c458138a7ccf-kube-api-access-9t4dw\") pod \"openstack-galera-0\" (UID: \"75d94fff-d912-4493-b9b9-c458138a7ccf\") " pod="openstack/openstack-galera-0" Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.968842 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/75d94fff-d912-4493-b9b9-c458138a7ccf-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"75d94fff-d912-4493-b9b9-c458138a7ccf\") " pod="openstack/openstack-galera-0" Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.968870 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75d94fff-d912-4493-b9b9-c458138a7ccf-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"75d94fff-d912-4493-b9b9-c458138a7ccf\") " pod="openstack/openstack-galera-0" Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.969579 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75d94fff-d912-4493-b9b9-c458138a7ccf-operator-scripts\") pod \"openstack-galera-0\" (UID: \"75d94fff-d912-4493-b9b9-c458138a7ccf\") " pod="openstack/openstack-galera-0" Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.969644 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"75d94fff-d912-4493-b9b9-c458138a7ccf\") " pod="openstack/openstack-galera-0" Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.969736 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/75d94fff-d912-4493-b9b9-c458138a7ccf-config-data-generated\") pod \"openstack-galera-0\" (UID: \"75d94fff-d912-4493-b9b9-c458138a7ccf\") " pod="openstack/openstack-galera-0" Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.971123 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/75d94fff-d912-4493-b9b9-c458138a7ccf-config-data-default\") pod \"openstack-galera-0\" (UID: \"75d94fff-d912-4493-b9b9-c458138a7ccf\") " pod="openstack/openstack-galera-0" Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.971790 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/75d94fff-d912-4493-b9b9-c458138a7ccf-config-data-generated\") pod \"openstack-galera-0\" (UID: \"75d94fff-d912-4493-b9b9-c458138a7ccf\") " pod="openstack/openstack-galera-0" Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.972142 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"75d94fff-d912-4493-b9b9-c458138a7ccf\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-galera-0" Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.973351 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/75d94fff-d912-4493-b9b9-c458138a7ccf-kolla-config\") pod \"openstack-galera-0\" (UID: \"75d94fff-d912-4493-b9b9-c458138a7ccf\") " pod="openstack/openstack-galera-0" Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.973887 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75d94fff-d912-4493-b9b9-c458138a7ccf-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"75d94fff-d912-4493-b9b9-c458138a7ccf\") " pod="openstack/openstack-galera-0" Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.975034 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75d94fff-d912-4493-b9b9-c458138a7ccf-operator-scripts\") pod \"openstack-galera-0\" (UID: \"75d94fff-d912-4493-b9b9-c458138a7ccf\") " pod="openstack/openstack-galera-0" Dec 06 16:01:01 crc kubenswrapper[4813]: I1206 16:01:01.981863 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/75d94fff-d912-4493-b9b9-c458138a7ccf-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"75d94fff-d912-4493-b9b9-c458138a7ccf\") " pod="openstack/openstack-galera-0" Dec 06 16:01:02 crc kubenswrapper[4813]: I1206 16:01:02.006364 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t4dw\" (UniqueName: \"kubernetes.io/projected/75d94fff-d912-4493-b9b9-c458138a7ccf-kube-api-access-9t4dw\") pod \"openstack-galera-0\" (UID: \"75d94fff-d912-4493-b9b9-c458138a7ccf\") " pod="openstack/openstack-galera-0" Dec 06 16:01:02 crc kubenswrapper[4813]: I1206 16:01:02.007829 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"75d94fff-d912-4493-b9b9-c458138a7ccf\") " pod="openstack/openstack-galera-0" Dec 06 16:01:02 crc kubenswrapper[4813]: I1206 16:01:02.069868 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 06 16:01:02 crc kubenswrapper[4813]: I1206 16:01:02.220937 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p2k89" podUID="d7c84964-2748-489e-9397-750792497b19" containerName="registry-server" containerID="cri-o://8010dc56575a8c0de633f5111b9c41543e7fa4be4d19dcdff3279ae3b42501a2" gracePeriod=2 Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.094532 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.097024 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.098416 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-qgf7h" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.100166 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.100333 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.100446 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.138466 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.187038 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgq6m\" (UniqueName: \"kubernetes.io/projected/bf7377f8-ef23-484a-b66b-90db1cd5b7cd-kube-api-access-jgq6m\") pod \"openstack-cell1-galera-0\" (UID: \"bf7377f8-ef23-484a-b66b-90db1cd5b7cd\") " pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.187082 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"bf7377f8-ef23-484a-b66b-90db1cd5b7cd\") " pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.187181 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bf7377f8-ef23-484a-b66b-90db1cd5b7cd-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"bf7377f8-ef23-484a-b66b-90db1cd5b7cd\") " pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.187226 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf7377f8-ef23-484a-b66b-90db1cd5b7cd-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"bf7377f8-ef23-484a-b66b-90db1cd5b7cd\") " pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.187275 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bf7377f8-ef23-484a-b66b-90db1cd5b7cd-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"bf7377f8-ef23-484a-b66b-90db1cd5b7cd\") " pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.187295 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf7377f8-ef23-484a-b66b-90db1cd5b7cd-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"bf7377f8-ef23-484a-b66b-90db1cd5b7cd\") " pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.187315 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/bf7377f8-ef23-484a-b66b-90db1cd5b7cd-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"bf7377f8-ef23-484a-b66b-90db1cd5b7cd\") " pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.187342 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/bf7377f8-ef23-484a-b66b-90db1cd5b7cd-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"bf7377f8-ef23-484a-b66b-90db1cd5b7cd\") " pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.251243 4813 generic.go:334] "Generic (PLEG): container finished" podID="d7c84964-2748-489e-9397-750792497b19" containerID="8010dc56575a8c0de633f5111b9c41543e7fa4be4d19dcdff3279ae3b42501a2" exitCode=0 Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.251899 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p2k89" event={"ID":"d7c84964-2748-489e-9397-750792497b19","Type":"ContainerDied","Data":"8010dc56575a8c0de633f5111b9c41543e7fa4be4d19dcdff3279ae3b42501a2"} Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.294046 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bf7377f8-ef23-484a-b66b-90db1cd5b7cd-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"bf7377f8-ef23-484a-b66b-90db1cd5b7cd\") " pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.294099 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf7377f8-ef23-484a-b66b-90db1cd5b7cd-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"bf7377f8-ef23-484a-b66b-90db1cd5b7cd\") " pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.294122 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/bf7377f8-ef23-484a-b66b-90db1cd5b7cd-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"bf7377f8-ef23-484a-b66b-90db1cd5b7cd\") " pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.294140 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/bf7377f8-ef23-484a-b66b-90db1cd5b7cd-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"bf7377f8-ef23-484a-b66b-90db1cd5b7cd\") " pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.294191 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgq6m\" (UniqueName: \"kubernetes.io/projected/bf7377f8-ef23-484a-b66b-90db1cd5b7cd-kube-api-access-jgq6m\") pod \"openstack-cell1-galera-0\" (UID: \"bf7377f8-ef23-484a-b66b-90db1cd5b7cd\") " pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.294217 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"bf7377f8-ef23-484a-b66b-90db1cd5b7cd\") " pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.294243 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bf7377f8-ef23-484a-b66b-90db1cd5b7cd-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"bf7377f8-ef23-484a-b66b-90db1cd5b7cd\") " pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.294318 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf7377f8-ef23-484a-b66b-90db1cd5b7cd-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"bf7377f8-ef23-484a-b66b-90db1cd5b7cd\") " pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.295463 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/bf7377f8-ef23-484a-b66b-90db1cd5b7cd-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"bf7377f8-ef23-484a-b66b-90db1cd5b7cd\") " pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.296208 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/bf7377f8-ef23-484a-b66b-90db1cd5b7cd-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"bf7377f8-ef23-484a-b66b-90db1cd5b7cd\") " pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.296292 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bf7377f8-ef23-484a-b66b-90db1cd5b7cd-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"bf7377f8-ef23-484a-b66b-90db1cd5b7cd\") " pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.297145 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bf7377f8-ef23-484a-b66b-90db1cd5b7cd-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"bf7377f8-ef23-484a-b66b-90db1cd5b7cd\") " pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.297296 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"bf7377f8-ef23-484a-b66b-90db1cd5b7cd\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.304951 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf7377f8-ef23-484a-b66b-90db1cd5b7cd-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"bf7377f8-ef23-484a-b66b-90db1cd5b7cd\") " pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.305396 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf7377f8-ef23-484a-b66b-90db1cd5b7cd-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"bf7377f8-ef23-484a-b66b-90db1cd5b7cd\") " pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.344706 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgq6m\" (UniqueName: \"kubernetes.io/projected/bf7377f8-ef23-484a-b66b-90db1cd5b7cd-kube-api-access-jgq6m\") pod \"openstack-cell1-galera-0\" (UID: \"bf7377f8-ef23-484a-b66b-90db1cd5b7cd\") " pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.354453 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"bf7377f8-ef23-484a-b66b-90db1cd5b7cd\") " pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.420418 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.539506 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.540410 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.548360 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.548490 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.548525 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-mvm4d" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.558639 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.716433 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c-config-data\") pod \"memcached-0\" (UID: \"9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c\") " pod="openstack/memcached-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.716486 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c-kolla-config\") pod \"memcached-0\" (UID: \"9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c\") " pod="openstack/memcached-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.716652 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c-combined-ca-bundle\") pod \"memcached-0\" (UID: \"9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c\") " pod="openstack/memcached-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.716717 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t74mk\" (UniqueName: \"kubernetes.io/projected/9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c-kube-api-access-t74mk\") pod \"memcached-0\" (UID: \"9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c\") " pod="openstack/memcached-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.716828 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c-memcached-tls-certs\") pod \"memcached-0\" (UID: \"9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c\") " pod="openstack/memcached-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.818626 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c-config-data\") pod \"memcached-0\" (UID: \"9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c\") " pod="openstack/memcached-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.818682 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c-kolla-config\") pod \"memcached-0\" (UID: \"9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c\") " pod="openstack/memcached-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.818708 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c-combined-ca-bundle\") pod \"memcached-0\" (UID: \"9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c\") " pod="openstack/memcached-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.818731 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t74mk\" (UniqueName: \"kubernetes.io/projected/9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c-kube-api-access-t74mk\") pod \"memcached-0\" (UID: \"9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c\") " pod="openstack/memcached-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.818762 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c-memcached-tls-certs\") pod \"memcached-0\" (UID: \"9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c\") " pod="openstack/memcached-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.819450 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c-kolla-config\") pod \"memcached-0\" (UID: \"9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c\") " pod="openstack/memcached-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.819965 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c-config-data\") pod \"memcached-0\" (UID: \"9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c\") " pod="openstack/memcached-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.835243 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t74mk\" (UniqueName: \"kubernetes.io/projected/9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c-kube-api-access-t74mk\") pod \"memcached-0\" (UID: \"9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c\") " pod="openstack/memcached-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.838719 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c-combined-ca-bundle\") pod \"memcached-0\" (UID: \"9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c\") " pod="openstack/memcached-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.839154 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c-memcached-tls-certs\") pod \"memcached-0\" (UID: \"9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c\") " pod="openstack/memcached-0" Dec 06 16:01:03 crc kubenswrapper[4813]: I1206 16:01:03.863759 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 06 16:01:04 crc kubenswrapper[4813]: I1206 16:01:04.762941 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-g8tcs" podUID="e9cd8622-0e78-400c-9de3-213fe41d1c1a" containerName="registry-server" probeResult="failure" output=< Dec 06 16:01:04 crc kubenswrapper[4813]: timeout: failed to connect service ":50051" within 1s Dec 06 16:01:04 crc kubenswrapper[4813]: > Dec 06 16:01:05 crc kubenswrapper[4813]: I1206 16:01:05.324076 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 16:01:05 crc kubenswrapper[4813]: I1206 16:01:05.331055 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 16:01:05 crc kubenswrapper[4813]: I1206 16:01:05.334571 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-db2m9" Dec 06 16:01:05 crc kubenswrapper[4813]: I1206 16:01:05.337581 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 16:01:05 crc kubenswrapper[4813]: I1206 16:01:05.440741 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wt7cv\" (UniqueName: \"kubernetes.io/projected/c9961e9d-0fa2-4619-9e19-9b2dfd4af40c-kube-api-access-wt7cv\") pod \"kube-state-metrics-0\" (UID: \"c9961e9d-0fa2-4619-9e19-9b2dfd4af40c\") " pod="openstack/kube-state-metrics-0" Dec 06 16:01:05 crc kubenswrapper[4813]: I1206 16:01:05.542060 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wt7cv\" (UniqueName: \"kubernetes.io/projected/c9961e9d-0fa2-4619-9e19-9b2dfd4af40c-kube-api-access-wt7cv\") pod \"kube-state-metrics-0\" (UID: \"c9961e9d-0fa2-4619-9e19-9b2dfd4af40c\") " pod="openstack/kube-state-metrics-0" Dec 06 16:01:05 crc kubenswrapper[4813]: I1206 16:01:05.558183 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wt7cv\" (UniqueName: \"kubernetes.io/projected/c9961e9d-0fa2-4619-9e19-9b2dfd4af40c-kube-api-access-wt7cv\") pod \"kube-state-metrics-0\" (UID: \"c9961e9d-0fa2-4619-9e19-9b2dfd4af40c\") " pod="openstack/kube-state-metrics-0" Dec 06 16:01:05 crc kubenswrapper[4813]: I1206 16:01:05.693875 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 16:01:09 crc kubenswrapper[4813]: E1206 16:01:09.884424 4813 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8010dc56575a8c0de633f5111b9c41543e7fa4be4d19dcdff3279ae3b42501a2 is running failed: container process not found" containerID="8010dc56575a8c0de633f5111b9c41543e7fa4be4d19dcdff3279ae3b42501a2" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 16:01:09 crc kubenswrapper[4813]: E1206 16:01:09.884950 4813 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8010dc56575a8c0de633f5111b9c41543e7fa4be4d19dcdff3279ae3b42501a2 is running failed: container process not found" containerID="8010dc56575a8c0de633f5111b9c41543e7fa4be4d19dcdff3279ae3b42501a2" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 16:01:09 crc kubenswrapper[4813]: E1206 16:01:09.885795 4813 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8010dc56575a8c0de633f5111b9c41543e7fa4be4d19dcdff3279ae3b42501a2 is running failed: container process not found" containerID="8010dc56575a8c0de633f5111b9c41543e7fa4be4d19dcdff3279ae3b42501a2" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 16:01:09 crc kubenswrapper[4813]: E1206 16:01:09.885841 4813 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8010dc56575a8c0de633f5111b9c41543e7fa4be4d19dcdff3279ae3b42501a2 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-p2k89" podUID="d7c84964-2748-489e-9397-750792497b19" containerName="registry-server" Dec 06 16:01:09 crc kubenswrapper[4813]: I1206 16:01:09.958777 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-bg6cb"] Dec 06 16:01:09 crc kubenswrapper[4813]: I1206 16:01:09.959753 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bg6cb" Dec 06 16:01:09 crc kubenswrapper[4813]: I1206 16:01:09.963297 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 06 16:01:09 crc kubenswrapper[4813]: I1206 16:01:09.963748 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 06 16:01:09 crc kubenswrapper[4813]: I1206 16:01:09.968797 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-7fhv5" Dec 06 16:01:09 crc kubenswrapper[4813]: I1206 16:01:09.978152 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-bg6cb"] Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.008219 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-mk4nr"] Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.009832 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-mk4nr" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.056580 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-mk4nr"] Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.136141 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba6825c1-2137-46d6-a695-6bdd53a13ec4-combined-ca-bundle\") pod \"ovn-controller-bg6cb\" (UID: \"ba6825c1-2137-46d6-a695-6bdd53a13ec4\") " pod="openstack/ovn-controller-bg6cb" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.136188 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/6b7cce06-d3db-47cd-a542-a77bfdc42f82-var-log\") pod \"ovn-controller-ovs-mk4nr\" (UID: \"6b7cce06-d3db-47cd-a542-a77bfdc42f82\") " pod="openstack/ovn-controller-ovs-mk4nr" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.136229 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvkps\" (UniqueName: \"kubernetes.io/projected/6b7cce06-d3db-47cd-a542-a77bfdc42f82-kube-api-access-zvkps\") pod \"ovn-controller-ovs-mk4nr\" (UID: \"6b7cce06-d3db-47cd-a542-a77bfdc42f82\") " pod="openstack/ovn-controller-ovs-mk4nr" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.136252 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6b7cce06-d3db-47cd-a542-a77bfdc42f82-var-run\") pod \"ovn-controller-ovs-mk4nr\" (UID: \"6b7cce06-d3db-47cd-a542-a77bfdc42f82\") " pod="openstack/ovn-controller-ovs-mk4nr" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.136334 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b7cce06-d3db-47cd-a542-a77bfdc42f82-scripts\") pod \"ovn-controller-ovs-mk4nr\" (UID: \"6b7cce06-d3db-47cd-a542-a77bfdc42f82\") " pod="openstack/ovn-controller-ovs-mk4nr" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.136350 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sr75\" (UniqueName: \"kubernetes.io/projected/ba6825c1-2137-46d6-a695-6bdd53a13ec4-kube-api-access-4sr75\") pod \"ovn-controller-bg6cb\" (UID: \"ba6825c1-2137-46d6-a695-6bdd53a13ec4\") " pod="openstack/ovn-controller-bg6cb" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.136462 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ba6825c1-2137-46d6-a695-6bdd53a13ec4-var-log-ovn\") pod \"ovn-controller-bg6cb\" (UID: \"ba6825c1-2137-46d6-a695-6bdd53a13ec4\") " pod="openstack/ovn-controller-bg6cb" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.136527 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba6825c1-2137-46d6-a695-6bdd53a13ec4-ovn-controller-tls-certs\") pod \"ovn-controller-bg6cb\" (UID: \"ba6825c1-2137-46d6-a695-6bdd53a13ec4\") " pod="openstack/ovn-controller-bg6cb" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.136614 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ba6825c1-2137-46d6-a695-6bdd53a13ec4-var-run-ovn\") pod \"ovn-controller-bg6cb\" (UID: \"ba6825c1-2137-46d6-a695-6bdd53a13ec4\") " pod="openstack/ovn-controller-bg6cb" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.136633 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ba6825c1-2137-46d6-a695-6bdd53a13ec4-scripts\") pod \"ovn-controller-bg6cb\" (UID: \"ba6825c1-2137-46d6-a695-6bdd53a13ec4\") " pod="openstack/ovn-controller-bg6cb" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.136678 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/6b7cce06-d3db-47cd-a542-a77bfdc42f82-etc-ovs\") pod \"ovn-controller-ovs-mk4nr\" (UID: \"6b7cce06-d3db-47cd-a542-a77bfdc42f82\") " pod="openstack/ovn-controller-ovs-mk4nr" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.136740 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ba6825c1-2137-46d6-a695-6bdd53a13ec4-var-run\") pod \"ovn-controller-bg6cb\" (UID: \"ba6825c1-2137-46d6-a695-6bdd53a13ec4\") " pod="openstack/ovn-controller-bg6cb" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.136803 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/6b7cce06-d3db-47cd-a542-a77bfdc42f82-var-lib\") pod \"ovn-controller-ovs-mk4nr\" (UID: \"6b7cce06-d3db-47cd-a542-a77bfdc42f82\") " pod="openstack/ovn-controller-ovs-mk4nr" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.161867 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p2k89" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.238106 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/6b7cce06-d3db-47cd-a542-a77bfdc42f82-var-log\") pod \"ovn-controller-ovs-mk4nr\" (UID: \"6b7cce06-d3db-47cd-a542-a77bfdc42f82\") " pod="openstack/ovn-controller-ovs-mk4nr" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.238183 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvkps\" (UniqueName: \"kubernetes.io/projected/6b7cce06-d3db-47cd-a542-a77bfdc42f82-kube-api-access-zvkps\") pod \"ovn-controller-ovs-mk4nr\" (UID: \"6b7cce06-d3db-47cd-a542-a77bfdc42f82\") " pod="openstack/ovn-controller-ovs-mk4nr" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.238208 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6b7cce06-d3db-47cd-a542-a77bfdc42f82-var-run\") pod \"ovn-controller-ovs-mk4nr\" (UID: \"6b7cce06-d3db-47cd-a542-a77bfdc42f82\") " pod="openstack/ovn-controller-ovs-mk4nr" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.238224 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b7cce06-d3db-47cd-a542-a77bfdc42f82-scripts\") pod \"ovn-controller-ovs-mk4nr\" (UID: \"6b7cce06-d3db-47cd-a542-a77bfdc42f82\") " pod="openstack/ovn-controller-ovs-mk4nr" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.238240 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sr75\" (UniqueName: \"kubernetes.io/projected/ba6825c1-2137-46d6-a695-6bdd53a13ec4-kube-api-access-4sr75\") pod \"ovn-controller-bg6cb\" (UID: \"ba6825c1-2137-46d6-a695-6bdd53a13ec4\") " pod="openstack/ovn-controller-bg6cb" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.238279 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ba6825c1-2137-46d6-a695-6bdd53a13ec4-var-log-ovn\") pod \"ovn-controller-bg6cb\" (UID: \"ba6825c1-2137-46d6-a695-6bdd53a13ec4\") " pod="openstack/ovn-controller-bg6cb" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.238300 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba6825c1-2137-46d6-a695-6bdd53a13ec4-ovn-controller-tls-certs\") pod \"ovn-controller-bg6cb\" (UID: \"ba6825c1-2137-46d6-a695-6bdd53a13ec4\") " pod="openstack/ovn-controller-bg6cb" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.238903 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ba6825c1-2137-46d6-a695-6bdd53a13ec4-var-run-ovn\") pod \"ovn-controller-bg6cb\" (UID: \"ba6825c1-2137-46d6-a695-6bdd53a13ec4\") " pod="openstack/ovn-controller-bg6cb" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.238933 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ba6825c1-2137-46d6-a695-6bdd53a13ec4-scripts\") pod \"ovn-controller-bg6cb\" (UID: \"ba6825c1-2137-46d6-a695-6bdd53a13ec4\") " pod="openstack/ovn-controller-bg6cb" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.238959 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/6b7cce06-d3db-47cd-a542-a77bfdc42f82-etc-ovs\") pod \"ovn-controller-ovs-mk4nr\" (UID: \"6b7cce06-d3db-47cd-a542-a77bfdc42f82\") " pod="openstack/ovn-controller-ovs-mk4nr" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.238974 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ba6825c1-2137-46d6-a695-6bdd53a13ec4-var-run\") pod \"ovn-controller-bg6cb\" (UID: \"ba6825c1-2137-46d6-a695-6bdd53a13ec4\") " pod="openstack/ovn-controller-bg6cb" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.239004 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/6b7cce06-d3db-47cd-a542-a77bfdc42f82-var-lib\") pod \"ovn-controller-ovs-mk4nr\" (UID: \"6b7cce06-d3db-47cd-a542-a77bfdc42f82\") " pod="openstack/ovn-controller-ovs-mk4nr" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.239032 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba6825c1-2137-46d6-a695-6bdd53a13ec4-combined-ca-bundle\") pod \"ovn-controller-bg6cb\" (UID: \"ba6825c1-2137-46d6-a695-6bdd53a13ec4\") " pod="openstack/ovn-controller-bg6cb" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.239142 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6b7cce06-d3db-47cd-a542-a77bfdc42f82-var-run\") pod \"ovn-controller-ovs-mk4nr\" (UID: \"6b7cce06-d3db-47cd-a542-a77bfdc42f82\") " pod="openstack/ovn-controller-ovs-mk4nr" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.239305 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/6b7cce06-d3db-47cd-a542-a77bfdc42f82-var-log\") pod \"ovn-controller-ovs-mk4nr\" (UID: \"6b7cce06-d3db-47cd-a542-a77bfdc42f82\") " pod="openstack/ovn-controller-ovs-mk4nr" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.239656 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ba6825c1-2137-46d6-a695-6bdd53a13ec4-var-run-ovn\") pod \"ovn-controller-bg6cb\" (UID: \"ba6825c1-2137-46d6-a695-6bdd53a13ec4\") " pod="openstack/ovn-controller-bg6cb" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.241306 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b7cce06-d3db-47cd-a542-a77bfdc42f82-scripts\") pod \"ovn-controller-ovs-mk4nr\" (UID: \"6b7cce06-d3db-47cd-a542-a77bfdc42f82\") " pod="openstack/ovn-controller-ovs-mk4nr" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.241532 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ba6825c1-2137-46d6-a695-6bdd53a13ec4-var-log-ovn\") pod \"ovn-controller-bg6cb\" (UID: \"ba6825c1-2137-46d6-a695-6bdd53a13ec4\") " pod="openstack/ovn-controller-bg6cb" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.241936 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ba6825c1-2137-46d6-a695-6bdd53a13ec4-var-run\") pod \"ovn-controller-bg6cb\" (UID: \"ba6825c1-2137-46d6-a695-6bdd53a13ec4\") " pod="openstack/ovn-controller-bg6cb" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.243685 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ba6825c1-2137-46d6-a695-6bdd53a13ec4-scripts\") pod \"ovn-controller-bg6cb\" (UID: \"ba6825c1-2137-46d6-a695-6bdd53a13ec4\") " pod="openstack/ovn-controller-bg6cb" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.244452 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/6b7cce06-d3db-47cd-a542-a77bfdc42f82-var-lib\") pod \"ovn-controller-ovs-mk4nr\" (UID: \"6b7cce06-d3db-47cd-a542-a77bfdc42f82\") " pod="openstack/ovn-controller-ovs-mk4nr" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.245234 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba6825c1-2137-46d6-a695-6bdd53a13ec4-combined-ca-bundle\") pod \"ovn-controller-bg6cb\" (UID: \"ba6825c1-2137-46d6-a695-6bdd53a13ec4\") " pod="openstack/ovn-controller-bg6cb" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.253486 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/6b7cce06-d3db-47cd-a542-a77bfdc42f82-etc-ovs\") pod \"ovn-controller-ovs-mk4nr\" (UID: \"6b7cce06-d3db-47cd-a542-a77bfdc42f82\") " pod="openstack/ovn-controller-ovs-mk4nr" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.257285 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvkps\" (UniqueName: \"kubernetes.io/projected/6b7cce06-d3db-47cd-a542-a77bfdc42f82-kube-api-access-zvkps\") pod \"ovn-controller-ovs-mk4nr\" (UID: \"6b7cce06-d3db-47cd-a542-a77bfdc42f82\") " pod="openstack/ovn-controller-ovs-mk4nr" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.267591 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sr75\" (UniqueName: \"kubernetes.io/projected/ba6825c1-2137-46d6-a695-6bdd53a13ec4-kube-api-access-4sr75\") pod \"ovn-controller-bg6cb\" (UID: \"ba6825c1-2137-46d6-a695-6bdd53a13ec4\") " pod="openstack/ovn-controller-bg6cb" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.269113 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba6825c1-2137-46d6-a695-6bdd53a13ec4-ovn-controller-tls-certs\") pod \"ovn-controller-bg6cb\" (UID: \"ba6825c1-2137-46d6-a695-6bdd53a13ec4\") " pod="openstack/ovn-controller-bg6cb" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.281050 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bg6cb" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.303524 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p2k89" event={"ID":"d7c84964-2748-489e-9397-750792497b19","Type":"ContainerDied","Data":"a091e7bd77170f359e5f4e037d779f45b3d0bec2b598d17d476f82b927f3ecb1"} Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.303576 4813 scope.go:117] "RemoveContainer" containerID="8010dc56575a8c0de633f5111b9c41543e7fa4be4d19dcdff3279ae3b42501a2" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.303728 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p2k89" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.340363 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-mk4nr" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.340658 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7c84964-2748-489e-9397-750792497b19-catalog-content\") pod \"d7c84964-2748-489e-9397-750792497b19\" (UID: \"d7c84964-2748-489e-9397-750792497b19\") " Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.340764 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7c84964-2748-489e-9397-750792497b19-utilities\") pod \"d7c84964-2748-489e-9397-750792497b19\" (UID: \"d7c84964-2748-489e-9397-750792497b19\") " Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.340790 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnhb6\" (UniqueName: \"kubernetes.io/projected/d7c84964-2748-489e-9397-750792497b19-kube-api-access-gnhb6\") pod \"d7c84964-2748-489e-9397-750792497b19\" (UID: \"d7c84964-2748-489e-9397-750792497b19\") " Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.342114 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7c84964-2748-489e-9397-750792497b19-utilities" (OuterVolumeSpecName: "utilities") pod "d7c84964-2748-489e-9397-750792497b19" (UID: "d7c84964-2748-489e-9397-750792497b19"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.363054 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7c84964-2748-489e-9397-750792497b19-kube-api-access-gnhb6" (OuterVolumeSpecName: "kube-api-access-gnhb6") pod "d7c84964-2748-489e-9397-750792497b19" (UID: "d7c84964-2748-489e-9397-750792497b19"). InnerVolumeSpecName "kube-api-access-gnhb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.393568 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7c84964-2748-489e-9397-750792497b19-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d7c84964-2748-489e-9397-750792497b19" (UID: "d7c84964-2748-489e-9397-750792497b19"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.443163 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7c84964-2748-489e-9397-750792497b19-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.443193 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7c84964-2748-489e-9397-750792497b19-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.443203 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnhb6\" (UniqueName: \"kubernetes.io/projected/d7c84964-2748-489e-9397-750792497b19-kube-api-access-gnhb6\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.648986 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p2k89"] Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.659495 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p2k89"] Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.844738 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 06 16:01:10 crc kubenswrapper[4813]: E1206 16:01:10.846794 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7c84964-2748-489e-9397-750792497b19" containerName="extract-utilities" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.846814 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7c84964-2748-489e-9397-750792497b19" containerName="extract-utilities" Dec 06 16:01:10 crc kubenswrapper[4813]: E1206 16:01:10.846869 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7c84964-2748-489e-9397-750792497b19" containerName="extract-content" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.846877 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7c84964-2748-489e-9397-750792497b19" containerName="extract-content" Dec 06 16:01:10 crc kubenswrapper[4813]: E1206 16:01:10.846889 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7c84964-2748-489e-9397-750792497b19" containerName="registry-server" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.846895 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7c84964-2748-489e-9397-750792497b19" containerName="registry-server" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.847046 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7c84964-2748-489e-9397-750792497b19" containerName="registry-server" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.847867 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.851202 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-2wsfv" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.851437 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.851548 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.854543 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.856947 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.874888 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.960070 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bh6cw\" (UniqueName: \"kubernetes.io/projected/c0a272bd-6444-4c90-acb0-fb653bf3b180-kube-api-access-bh6cw\") pod \"ovsdbserver-nb-0\" (UID: \"c0a272bd-6444-4c90-acb0-fb653bf3b180\") " pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.960120 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c0a272bd-6444-4c90-acb0-fb653bf3b180-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c0a272bd-6444-4c90-acb0-fb653bf3b180\") " pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.960157 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c0a272bd-6444-4c90-acb0-fb653bf3b180\") " pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.960181 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0a272bd-6444-4c90-acb0-fb653bf3b180-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c0a272bd-6444-4c90-acb0-fb653bf3b180\") " pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.960244 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0a272bd-6444-4c90-acb0-fb653bf3b180-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c0a272bd-6444-4c90-acb0-fb653bf3b180\") " pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.960303 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0a272bd-6444-4c90-acb0-fb653bf3b180-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c0a272bd-6444-4c90-acb0-fb653bf3b180\") " pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.960381 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c0a272bd-6444-4c90-acb0-fb653bf3b180-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c0a272bd-6444-4c90-acb0-fb653bf3b180\") " pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:10 crc kubenswrapper[4813]: I1206 16:01:10.960512 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0a272bd-6444-4c90-acb0-fb653bf3b180-config\") pod \"ovsdbserver-nb-0\" (UID: \"c0a272bd-6444-4c90-acb0-fb653bf3b180\") " pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:11 crc kubenswrapper[4813]: I1206 16:01:11.061625 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0a272bd-6444-4c90-acb0-fb653bf3b180-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c0a272bd-6444-4c90-acb0-fb653bf3b180\") " pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:11 crc kubenswrapper[4813]: I1206 16:01:11.061683 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c0a272bd-6444-4c90-acb0-fb653bf3b180-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c0a272bd-6444-4c90-acb0-fb653bf3b180\") " pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:11 crc kubenswrapper[4813]: I1206 16:01:11.061746 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0a272bd-6444-4c90-acb0-fb653bf3b180-config\") pod \"ovsdbserver-nb-0\" (UID: \"c0a272bd-6444-4c90-acb0-fb653bf3b180\") " pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:11 crc kubenswrapper[4813]: I1206 16:01:11.061770 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bh6cw\" (UniqueName: \"kubernetes.io/projected/c0a272bd-6444-4c90-acb0-fb653bf3b180-kube-api-access-bh6cw\") pod \"ovsdbserver-nb-0\" (UID: \"c0a272bd-6444-4c90-acb0-fb653bf3b180\") " pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:11 crc kubenswrapper[4813]: I1206 16:01:11.061789 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c0a272bd-6444-4c90-acb0-fb653bf3b180-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c0a272bd-6444-4c90-acb0-fb653bf3b180\") " pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:11 crc kubenswrapper[4813]: I1206 16:01:11.061812 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c0a272bd-6444-4c90-acb0-fb653bf3b180\") " pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:11 crc kubenswrapper[4813]: I1206 16:01:11.061827 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0a272bd-6444-4c90-acb0-fb653bf3b180-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c0a272bd-6444-4c90-acb0-fb653bf3b180\") " pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:11 crc kubenswrapper[4813]: I1206 16:01:11.061870 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0a272bd-6444-4c90-acb0-fb653bf3b180-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c0a272bd-6444-4c90-acb0-fb653bf3b180\") " pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:11 crc kubenswrapper[4813]: I1206 16:01:11.063338 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c0a272bd-6444-4c90-acb0-fb653bf3b180-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c0a272bd-6444-4c90-acb0-fb653bf3b180\") " pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:11 crc kubenswrapper[4813]: I1206 16:01:11.064031 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0a272bd-6444-4c90-acb0-fb653bf3b180-config\") pod \"ovsdbserver-nb-0\" (UID: \"c0a272bd-6444-4c90-acb0-fb653bf3b180\") " pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:11 crc kubenswrapper[4813]: I1206 16:01:11.065026 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c0a272bd-6444-4c90-acb0-fb653bf3b180\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:11 crc kubenswrapper[4813]: I1206 16:01:11.065394 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c0a272bd-6444-4c90-acb0-fb653bf3b180-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c0a272bd-6444-4c90-acb0-fb653bf3b180\") " pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:11 crc kubenswrapper[4813]: I1206 16:01:11.069924 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0a272bd-6444-4c90-acb0-fb653bf3b180-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c0a272bd-6444-4c90-acb0-fb653bf3b180\") " pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:11 crc kubenswrapper[4813]: I1206 16:01:11.070611 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0a272bd-6444-4c90-acb0-fb653bf3b180-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c0a272bd-6444-4c90-acb0-fb653bf3b180\") " pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:11 crc kubenswrapper[4813]: I1206 16:01:11.071628 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0a272bd-6444-4c90-acb0-fb653bf3b180-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c0a272bd-6444-4c90-acb0-fb653bf3b180\") " pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:11 crc kubenswrapper[4813]: I1206 16:01:11.076949 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bh6cw\" (UniqueName: \"kubernetes.io/projected/c0a272bd-6444-4c90-acb0-fb653bf3b180-kube-api-access-bh6cw\") pod \"ovsdbserver-nb-0\" (UID: \"c0a272bd-6444-4c90-acb0-fb653bf3b180\") " pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:11 crc kubenswrapper[4813]: I1206 16:01:11.083621 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c0a272bd-6444-4c90-acb0-fb653bf3b180\") " pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:11 crc kubenswrapper[4813]: I1206 16:01:11.177175 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:12 crc kubenswrapper[4813]: I1206 16:01:12.494387 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7c84964-2748-489e-9397-750792497b19" path="/var/lib/kubelet/pods/d7c84964-2748-489e-9397-750792497b19/volumes" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.250053 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.252492 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.254320 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.254689 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-xp7ck" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.256414 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.256576 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.260756 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.407850 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420\") " pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.407906 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420\") " pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.407961 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420-config\") pod \"ovsdbserver-sb-0\" (UID: \"7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420\") " pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.407991 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zm2kl\" (UniqueName: \"kubernetes.io/projected/7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420-kube-api-access-zm2kl\") pod \"ovsdbserver-sb-0\" (UID: \"7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420\") " pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.408017 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420\") " pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.408059 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420\") " pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.408118 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420\") " pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.408146 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420\") " pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.509572 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420-config\") pod \"ovsdbserver-sb-0\" (UID: \"7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420\") " pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.509689 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zm2kl\" (UniqueName: \"kubernetes.io/projected/7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420-kube-api-access-zm2kl\") pod \"ovsdbserver-sb-0\" (UID: \"7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420\") " pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.509727 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420\") " pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.509770 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420\") " pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.509809 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420\") " pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.509869 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420\") " pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.509925 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420\") " pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.509979 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420\") " pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.512814 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.514451 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420\") " pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.519422 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420-config\") pod \"ovsdbserver-sb-0\" (UID: \"7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420\") " pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.519842 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420\") " pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.521025 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420\") " pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.522940 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420\") " pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.534413 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420\") " pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.540921 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zm2kl\" (UniqueName: \"kubernetes.io/projected/7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420-kube-api-access-zm2kl\") pod \"ovsdbserver-sb-0\" (UID: \"7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420\") " pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.558527 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420\") " pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.567750 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.732439 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-g8tcs" Dec 06 16:01:13 crc kubenswrapper[4813]: I1206 16:01:13.782178 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-g8tcs" Dec 06 16:01:14 crc kubenswrapper[4813]: I1206 16:01:14.555082 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g8tcs"] Dec 06 16:01:15 crc kubenswrapper[4813]: I1206 16:01:15.340302 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-g8tcs" podUID="e9cd8622-0e78-400c-9de3-213fe41d1c1a" containerName="registry-server" containerID="cri-o://a51bf28dba637d2e63af3ea9134313fcfecad4d56995810e2c33cddbd6b62283" gracePeriod=2 Dec 06 16:01:16 crc kubenswrapper[4813]: E1206 16:01:16.254002 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 06 16:01:16 crc kubenswrapper[4813]: E1206 16:01:16.254143 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ktsw7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-nng9s_openstack(ee786ef0-6a58-4266-ad9e-93a97b9d4cbf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 16:01:16 crc kubenswrapper[4813]: E1206 16:01:16.256054 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-nng9s" podUID="ee786ef0-6a58-4266-ad9e-93a97b9d4cbf" Dec 06 16:01:16 crc kubenswrapper[4813]: E1206 16:01:16.308386 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 06 16:01:16 crc kubenswrapper[4813]: E1206 16:01:16.309058 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bb49c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-whjs8_openstack(7e0d82a4-b89e-4c97-9038-e1554d04963b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 16:01:16 crc kubenswrapper[4813]: E1206 16:01:16.310402 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-whjs8" podUID="7e0d82a4-b89e-4c97-9038-e1554d04963b" Dec 06 16:01:16 crc kubenswrapper[4813]: I1206 16:01:16.359363 4813 scope.go:117] "RemoveContainer" containerID="38ed1e7a29dae2d294373293ba666fce3ea4a33c96e8e130f8a1cb866504b1c6" Dec 06 16:01:16 crc kubenswrapper[4813]: I1206 16:01:16.407433 4813 generic.go:334] "Generic (PLEG): container finished" podID="e9cd8622-0e78-400c-9de3-213fe41d1c1a" containerID="a51bf28dba637d2e63af3ea9134313fcfecad4d56995810e2c33cddbd6b62283" exitCode=0 Dec 06 16:01:16 crc kubenswrapper[4813]: I1206 16:01:16.407525 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8tcs" event={"ID":"e9cd8622-0e78-400c-9de3-213fe41d1c1a","Type":"ContainerDied","Data":"a51bf28dba637d2e63af3ea9134313fcfecad4d56995810e2c33cddbd6b62283"} Dec 06 16:01:16 crc kubenswrapper[4813]: I1206 16:01:16.592749 4813 scope.go:117] "RemoveContainer" containerID="cad225f5f92c252eb9a79a6b20de04be0ee9c7ba0dbb523d072c16933bc22b23" Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.422383 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-whjs8" event={"ID":"7e0d82a4-b89e-4c97-9038-e1554d04963b","Type":"ContainerDied","Data":"a1d1146607cb64ac90ee024db2a392925639d246fe6142bd69f43ddcffe8dee6"} Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.422824 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1d1146607cb64ac90ee024db2a392925639d246fe6142bd69f43ddcffe8dee6" Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.448324 4813 generic.go:334] "Generic (PLEG): container finished" podID="00d42d29-5c9b-4585-be1b-51077e528837" containerID="92ab4225b77064b8f8a05203d2ce0a3065fdb43b3f7cf55fb3addb1b45a5805e" exitCode=0 Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.448445 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-2fqgs" event={"ID":"00d42d29-5c9b-4585-be1b-51077e528837","Type":"ContainerDied","Data":"92ab4225b77064b8f8a05203d2ce0a3065fdb43b3f7cf55fb3addb1b45a5805e"} Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.479050 4813 generic.go:334] "Generic (PLEG): container finished" podID="fb685be0-8427-43c2-ac4e-262c73714b5f" containerID="3165b9a67a884f501a9e9e0fba148fb0f0151d101da7a4e0efdbdf19fac09b04" exitCode=0 Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.479111 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-6tgdq" event={"ID":"fb685be0-8427-43c2-ac4e-262c73714b5f","Type":"ContainerDied","Data":"3165b9a67a884f501a9e9e0fba148fb0f0151d101da7a4e0efdbdf19fac09b04"} Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.495829 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lwvb" event={"ID":"1f91a370-ae93-4992-95f9-ebb72180adec","Type":"ContainerStarted","Data":"3c58fba782bd161198ed1d4703b75ac9847279c921a8d696fdca53e276030fa7"} Dec 06 16:01:17 crc kubenswrapper[4813]: W1206 16:01:17.530088 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9fab2eee_6b9d_4fd2_b3cc_0468fed71b2c.slice/crio-94aeafe4aa3df86fe5924eda7d6b324831c57d4f1d95242c0fc492fb034a7219 WatchSource:0}: Error finding container 94aeafe4aa3df86fe5924eda7d6b324831c57d4f1d95242c0fc492fb034a7219: Status 404 returned error can't find the container with id 94aeafe4aa3df86fe5924eda7d6b324831c57d4f1d95242c0fc492fb034a7219 Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.627286 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.656813 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-whjs8" Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.670047 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-nng9s" Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.674522 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.707468 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bb49c\" (UniqueName: \"kubernetes.io/projected/7e0d82a4-b89e-4c97-9038-e1554d04963b-kube-api-access-bb49c\") pod \"7e0d82a4-b89e-4c97-9038-e1554d04963b\" (UID: \"7e0d82a4-b89e-4c97-9038-e1554d04963b\") " Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.707505 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e0d82a4-b89e-4c97-9038-e1554d04963b-dns-svc\") pod \"7e0d82a4-b89e-4c97-9038-e1554d04963b\" (UID: \"7e0d82a4-b89e-4c97-9038-e1554d04963b\") " Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.707584 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e0d82a4-b89e-4c97-9038-e1554d04963b-config\") pod \"7e0d82a4-b89e-4c97-9038-e1554d04963b\" (UID: \"7e0d82a4-b89e-4c97-9038-e1554d04963b\") " Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.708283 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e0d82a4-b89e-4c97-9038-e1554d04963b-config" (OuterVolumeSpecName: "config") pod "7e0d82a4-b89e-4c97-9038-e1554d04963b" (UID: "7e0d82a4-b89e-4c97-9038-e1554d04963b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.710285 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e0d82a4-b89e-4c97-9038-e1554d04963b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7e0d82a4-b89e-4c97-9038-e1554d04963b" (UID: "7e0d82a4-b89e-4c97-9038-e1554d04963b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.727559 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e0d82a4-b89e-4c97-9038-e1554d04963b-kube-api-access-bb49c" (OuterVolumeSpecName: "kube-api-access-bb49c") pod "7e0d82a4-b89e-4c97-9038-e1554d04963b" (UID: "7e0d82a4-b89e-4c97-9038-e1554d04963b"). InnerVolumeSpecName "kube-api-access-bb49c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.734610 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4lwvb" podStartSLOduration=3.448057519 podStartE2EDuration="22.734595762s" podCreationTimestamp="2025-12-06 16:00:55 +0000 UTC" firstStartedPulling="2025-12-06 16:00:57.143850779 +0000 UTC m=+897.034730355" lastFinishedPulling="2025-12-06 16:01:16.430389032 +0000 UTC m=+916.321268598" observedRunningTime="2025-12-06 16:01:17.589865772 +0000 UTC m=+917.480745348" watchObservedRunningTime="2025-12-06 16:01:17.734595762 +0000 UTC m=+917.625475338" Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.769518 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.777913 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.808503 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.809187 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee786ef0-6a58-4266-ad9e-93a97b9d4cbf-config\") pod \"ee786ef0-6a58-4266-ad9e-93a97b9d4cbf\" (UID: \"ee786ef0-6a58-4266-ad9e-93a97b9d4cbf\") " Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.809236 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktsw7\" (UniqueName: \"kubernetes.io/projected/ee786ef0-6a58-4266-ad9e-93a97b9d4cbf-kube-api-access-ktsw7\") pod \"ee786ef0-6a58-4266-ad9e-93a97b9d4cbf\" (UID: \"ee786ef0-6a58-4266-ad9e-93a97b9d4cbf\") " Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.809695 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bb49c\" (UniqueName: \"kubernetes.io/projected/7e0d82a4-b89e-4c97-9038-e1554d04963b-kube-api-access-bb49c\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.809709 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e0d82a4-b89e-4c97-9038-e1554d04963b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.809717 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e0d82a4-b89e-4c97-9038-e1554d04963b-config\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:17 crc kubenswrapper[4813]: W1206 16:01:17.808196 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75d94fff_d912_4493_b9b9_c458138a7ccf.slice/crio-43c35bbc013cdd56f186fb5d69cb6a20f3108663d7b11d96d1d07c313ca733c2 WatchSource:0}: Error finding container 43c35bbc013cdd56f186fb5d69cb6a20f3108663d7b11d96d1d07c313ca733c2: Status 404 returned error can't find the container with id 43c35bbc013cdd56f186fb5d69cb6a20f3108663d7b11d96d1d07c313ca733c2 Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.811128 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee786ef0-6a58-4266-ad9e-93a97b9d4cbf-config" (OuterVolumeSpecName: "config") pod "ee786ef0-6a58-4266-ad9e-93a97b9d4cbf" (UID: "ee786ef0-6a58-4266-ad9e-93a97b9d4cbf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.811954 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee786ef0-6a58-4266-ad9e-93a97b9d4cbf-kube-api-access-ktsw7" (OuterVolumeSpecName: "kube-api-access-ktsw7") pod "ee786ef0-6a58-4266-ad9e-93a97b9d4cbf" (UID: "ee786ef0-6a58-4266-ad9e-93a97b9d4cbf"). InnerVolumeSpecName "kube-api-access-ktsw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.820054 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.832422 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8tcs" Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.911419 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9cd8622-0e78-400c-9de3-213fe41d1c1a-catalog-content\") pod \"e9cd8622-0e78-400c-9de3-213fe41d1c1a\" (UID: \"e9cd8622-0e78-400c-9de3-213fe41d1c1a\") " Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.911530 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9vgp\" (UniqueName: \"kubernetes.io/projected/e9cd8622-0e78-400c-9de3-213fe41d1c1a-kube-api-access-n9vgp\") pod \"e9cd8622-0e78-400c-9de3-213fe41d1c1a\" (UID: \"e9cd8622-0e78-400c-9de3-213fe41d1c1a\") " Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.911629 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9cd8622-0e78-400c-9de3-213fe41d1c1a-utilities\") pod \"e9cd8622-0e78-400c-9de3-213fe41d1c1a\" (UID: \"e9cd8622-0e78-400c-9de3-213fe41d1c1a\") " Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.911942 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee786ef0-6a58-4266-ad9e-93a97b9d4cbf-config\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.911954 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktsw7\" (UniqueName: \"kubernetes.io/projected/ee786ef0-6a58-4266-ad9e-93a97b9d4cbf-kube-api-access-ktsw7\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.912807 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9cd8622-0e78-400c-9de3-213fe41d1c1a-utilities" (OuterVolumeSpecName: "utilities") pod "e9cd8622-0e78-400c-9de3-213fe41d1c1a" (UID: "e9cd8622-0e78-400c-9de3-213fe41d1c1a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:01:17 crc kubenswrapper[4813]: I1206 16:01:17.922040 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9cd8622-0e78-400c-9de3-213fe41d1c1a-kube-api-access-n9vgp" (OuterVolumeSpecName: "kube-api-access-n9vgp") pod "e9cd8622-0e78-400c-9de3-213fe41d1c1a" (UID: "e9cd8622-0e78-400c-9de3-213fe41d1c1a"). InnerVolumeSpecName "kube-api-access-n9vgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:01:18 crc kubenswrapper[4813]: I1206 16:01:18.013402 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9vgp\" (UniqueName: \"kubernetes.io/projected/e9cd8622-0e78-400c-9de3-213fe41d1c1a-kube-api-access-n9vgp\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:18 crc kubenswrapper[4813]: I1206 16:01:18.013427 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9cd8622-0e78-400c-9de3-213fe41d1c1a-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:18 crc kubenswrapper[4813]: E1206 16:01:18.041675 4813 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Dec 06 16:01:18 crc kubenswrapper[4813]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/00d42d29-5c9b-4585-be1b-51077e528837/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 06 16:01:18 crc kubenswrapper[4813]: > podSandboxID="495e4ccbc38c31136a94ca11755afb0728a50af7fc99d86c23845fc9cf0b6dfd" Dec 06 16:01:18 crc kubenswrapper[4813]: E1206 16:01:18.041813 4813 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 06 16:01:18 crc kubenswrapper[4813]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ggxdn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5ccc8479f9-2fqgs_openstack(00d42d29-5c9b-4585-be1b-51077e528837): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/00d42d29-5c9b-4585-be1b-51077e528837/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 06 16:01:18 crc kubenswrapper[4813]: > logger="UnhandledError" Dec 06 16:01:18 crc kubenswrapper[4813]: E1206 16:01:18.044374 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/00d42d29-5c9b-4585-be1b-51077e528837/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-5ccc8479f9-2fqgs" podUID="00d42d29-5c9b-4585-be1b-51077e528837" Dec 06 16:01:18 crc kubenswrapper[4813]: I1206 16:01:18.054555 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9cd8622-0e78-400c-9de3-213fe41d1c1a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e9cd8622-0e78-400c-9de3-213fe41d1c1a" (UID: "e9cd8622-0e78-400c-9de3-213fe41d1c1a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:01:18 crc kubenswrapper[4813]: I1206 16:01:18.114390 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9cd8622-0e78-400c-9de3-213fe41d1c1a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:18 crc kubenswrapper[4813]: I1206 16:01:18.120936 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-bg6cb"] Dec 06 16:01:18 crc kubenswrapper[4813]: I1206 16:01:18.338230 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-mk4nr"] Dec 06 16:01:18 crc kubenswrapper[4813]: W1206 16:01:18.357221 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b7cce06_d3db_47cd_a542_a77bfdc42f82.slice/crio-2b202234ebf6d919252e24a5889ccdcc68d81b4323a01e6380da54108a175937 WatchSource:0}: Error finding container 2b202234ebf6d919252e24a5889ccdcc68d81b4323a01e6380da54108a175937: Status 404 returned error can't find the container with id 2b202234ebf6d919252e24a5889ccdcc68d81b4323a01e6380da54108a175937 Dec 06 16:01:18 crc kubenswrapper[4813]: I1206 16:01:18.517313 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"75d94fff-d912-4493-b9b9-c458138a7ccf","Type":"ContainerStarted","Data":"43c35bbc013cdd56f186fb5d69cb6a20f3108663d7b11d96d1d07c313ca733c2"} Dec 06 16:01:18 crc kubenswrapper[4813]: I1206 16:01:18.517525 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-bg6cb" event={"ID":"ba6825c1-2137-46d6-a695-6bdd53a13ec4","Type":"ContainerStarted","Data":"4b0cd32e6ae8137a5101025f22b4effa1b8f80df41a0c24a5f876c50cab7571e"} Dec 06 16:01:18 crc kubenswrapper[4813]: I1206 16:01:18.519211 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19","Type":"ContainerStarted","Data":"a58091e3567d12b58c4bbae8d972435c43fedb3d6e70bdf69419dba603eaedea"} Dec 06 16:01:18 crc kubenswrapper[4813]: I1206 16:01:18.523052 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-6tgdq" event={"ID":"fb685be0-8427-43c2-ac4e-262c73714b5f","Type":"ContainerStarted","Data":"13d1fa7406788ff72d6bcf5296b8907beca0f71d1196701e3d60205544a650e9"} Dec 06 16:01:18 crc kubenswrapper[4813]: I1206 16:01:18.523440 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-6tgdq" Dec 06 16:01:18 crc kubenswrapper[4813]: I1206 16:01:18.527605 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"bf7377f8-ef23-484a-b66b-90db1cd5b7cd","Type":"ContainerStarted","Data":"a96bb45caa02435838306919360a72f5a4d478903cc4e66efa83ead45f1f62ec"} Dec 06 16:01:18 crc kubenswrapper[4813]: I1206 16:01:18.529295 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2172dfb5-c0a6-4525-a326-dd9736eb5233","Type":"ContainerStarted","Data":"46805f68b4ee42a603c02323712604516589877604fd0908587d6bc9a4a4616f"} Dec 06 16:01:18 crc kubenswrapper[4813]: I1206 16:01:18.530920 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mk4nr" event={"ID":"6b7cce06-d3db-47cd-a542-a77bfdc42f82","Type":"ContainerStarted","Data":"2b202234ebf6d919252e24a5889ccdcc68d81b4323a01e6380da54108a175937"} Dec 06 16:01:18 crc kubenswrapper[4813]: I1206 16:01:18.532858 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c9961e9d-0fa2-4619-9e19-9b2dfd4af40c","Type":"ContainerStarted","Data":"6d6e3a94cb9511ce35608cbdc770d2eed67613433e2e2b89e2eb113161b936d4"} Dec 06 16:01:18 crc kubenswrapper[4813]: I1206 16:01:18.535095 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-nng9s" event={"ID":"ee786ef0-6a58-4266-ad9e-93a97b9d4cbf","Type":"ContainerDied","Data":"160bcf4587ef9b3e65ef92be86c9643f98cbe7adaeb14eeb91479c3979e79a82"} Dec 06 16:01:18 crc kubenswrapper[4813]: I1206 16:01:18.535165 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-nng9s" Dec 06 16:01:18 crc kubenswrapper[4813]: I1206 16:01:18.544883 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-6tgdq" podStartSLOduration=2.976803814 podStartE2EDuration="19.544863359s" podCreationTimestamp="2025-12-06 16:00:59 +0000 UTC" firstStartedPulling="2025-12-06 16:01:00.071072541 +0000 UTC m=+899.961952117" lastFinishedPulling="2025-12-06 16:01:16.639132086 +0000 UTC m=+916.530011662" observedRunningTime="2025-12-06 16:01:18.538819449 +0000 UTC m=+918.429699025" watchObservedRunningTime="2025-12-06 16:01:18.544863359 +0000 UTC m=+918.435742935" Dec 06 16:01:18 crc kubenswrapper[4813]: I1206 16:01:18.560150 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8tcs" event={"ID":"e9cd8622-0e78-400c-9de3-213fe41d1c1a","Type":"ContainerDied","Data":"4d85bf2e9e73accdca27189247fd184fe4fbd461b46a0b6fe18d7c8c043929e3"} Dec 06 16:01:18 crc kubenswrapper[4813]: I1206 16:01:18.560197 4813 scope.go:117] "RemoveContainer" containerID="a51bf28dba637d2e63af3ea9134313fcfecad4d56995810e2c33cddbd6b62283" Dec 06 16:01:18 crc kubenswrapper[4813]: I1206 16:01:18.560354 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8tcs" Dec 06 16:01:18 crc kubenswrapper[4813]: I1206 16:01:18.563492 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-whjs8" Dec 06 16:01:18 crc kubenswrapper[4813]: I1206 16:01:18.563252 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c","Type":"ContainerStarted","Data":"94aeafe4aa3df86fe5924eda7d6b324831c57d4f1d95242c0fc492fb034a7219"} Dec 06 16:01:18 crc kubenswrapper[4813]: I1206 16:01:18.639984 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-whjs8"] Dec 06 16:01:18 crc kubenswrapper[4813]: I1206 16:01:18.646896 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-whjs8"] Dec 06 16:01:18 crc kubenswrapper[4813]: I1206 16:01:18.674560 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g8tcs"] Dec 06 16:01:18 crc kubenswrapper[4813]: I1206 16:01:18.716469 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-g8tcs"] Dec 06 16:01:18 crc kubenswrapper[4813]: I1206 16:01:18.749422 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-nng9s"] Dec 06 16:01:18 crc kubenswrapper[4813]: I1206 16:01:18.761568 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-nng9s"] Dec 06 16:01:19 crc kubenswrapper[4813]: I1206 16:01:19.021743 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 06 16:01:19 crc kubenswrapper[4813]: I1206 16:01:19.119245 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 06 16:01:19 crc kubenswrapper[4813]: W1206 16:01:19.618137 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7cce2bf2_3398_47a8_8d0d_ae8ef8ebd420.slice/crio-e5aaaec77ae07a0fc21e07db3fd296c59ebc8919d0576d55144cec63d1939848 WatchSource:0}: Error finding container e5aaaec77ae07a0fc21e07db3fd296c59ebc8919d0576d55144cec63d1939848: Status 404 returned error can't find the container with id e5aaaec77ae07a0fc21e07db3fd296c59ebc8919d0576d55144cec63d1939848 Dec 06 16:01:20 crc kubenswrapper[4813]: I1206 16:01:20.504792 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e0d82a4-b89e-4c97-9038-e1554d04963b" path="/var/lib/kubelet/pods/7e0d82a4-b89e-4c97-9038-e1554d04963b/volumes" Dec 06 16:01:20 crc kubenswrapper[4813]: I1206 16:01:20.505453 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9cd8622-0e78-400c-9de3-213fe41d1c1a" path="/var/lib/kubelet/pods/e9cd8622-0e78-400c-9de3-213fe41d1c1a/volumes" Dec 06 16:01:20 crc kubenswrapper[4813]: I1206 16:01:20.506197 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee786ef0-6a58-4266-ad9e-93a97b9d4cbf" path="/var/lib/kubelet/pods/ee786ef0-6a58-4266-ad9e-93a97b9d4cbf/volumes" Dec 06 16:01:20 crc kubenswrapper[4813]: I1206 16:01:20.600091 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420","Type":"ContainerStarted","Data":"e5aaaec77ae07a0fc21e07db3fd296c59ebc8919d0576d55144cec63d1939848"} Dec 06 16:01:21 crc kubenswrapper[4813]: I1206 16:01:21.609607 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c0a272bd-6444-4c90-acb0-fb653bf3b180","Type":"ContainerStarted","Data":"cd2991e00537e1a596bcab07d27aa0ff5098a4e16d1158a3f9d87b30a6d1d40f"} Dec 06 16:01:22 crc kubenswrapper[4813]: I1206 16:01:22.556500 4813 scope.go:117] "RemoveContainer" containerID="4c2c8620754fe18f4e2b2a097f9aabf2003489a56a3a8ae3aea6ca46c4f3f211" Dec 06 16:01:24 crc kubenswrapper[4813]: I1206 16:01:24.529682 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-6tgdq" Dec 06 16:01:24 crc kubenswrapper[4813]: I1206 16:01:24.616877 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-2fqgs"] Dec 06 16:01:26 crc kubenswrapper[4813]: I1206 16:01:26.063210 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4lwvb" Dec 06 16:01:26 crc kubenswrapper[4813]: I1206 16:01:26.063997 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4lwvb" Dec 06 16:01:26 crc kubenswrapper[4813]: I1206 16:01:26.104171 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4lwvb" Dec 06 16:01:26 crc kubenswrapper[4813]: I1206 16:01:26.689995 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4lwvb" Dec 06 16:01:26 crc kubenswrapper[4813]: I1206 16:01:26.958591 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4lwvb"] Dec 06 16:01:28 crc kubenswrapper[4813]: I1206 16:01:28.392739 4813 scope.go:117] "RemoveContainer" containerID="fb08cadb487a1b92cd6fee118777a85510e021c071df300f847f8b323495e72b" Dec 06 16:01:28 crc kubenswrapper[4813]: I1206 16:01:28.667916 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4lwvb" podUID="1f91a370-ae93-4992-95f9-ebb72180adec" containerName="registry-server" containerID="cri-o://3c58fba782bd161198ed1d4703b75ac9847279c921a8d696fdca53e276030fa7" gracePeriod=2 Dec 06 16:01:29 crc kubenswrapper[4813]: I1206 16:01:29.690347 4813 generic.go:334] "Generic (PLEG): container finished" podID="1f91a370-ae93-4992-95f9-ebb72180adec" containerID="3c58fba782bd161198ed1d4703b75ac9847279c921a8d696fdca53e276030fa7" exitCode=0 Dec 06 16:01:29 crc kubenswrapper[4813]: I1206 16:01:29.690862 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lwvb" event={"ID":"1f91a370-ae93-4992-95f9-ebb72180adec","Type":"ContainerDied","Data":"3c58fba782bd161198ed1d4703b75ac9847279c921a8d696fdca53e276030fa7"} Dec 06 16:01:29 crc kubenswrapper[4813]: I1206 16:01:29.839491 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4lwvb" Dec 06 16:01:29 crc kubenswrapper[4813]: I1206 16:01:29.984015 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f91a370-ae93-4992-95f9-ebb72180adec-catalog-content\") pod \"1f91a370-ae93-4992-95f9-ebb72180adec\" (UID: \"1f91a370-ae93-4992-95f9-ebb72180adec\") " Dec 06 16:01:29 crc kubenswrapper[4813]: I1206 16:01:29.984054 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f91a370-ae93-4992-95f9-ebb72180adec-utilities\") pod \"1f91a370-ae93-4992-95f9-ebb72180adec\" (UID: \"1f91a370-ae93-4992-95f9-ebb72180adec\") " Dec 06 16:01:29 crc kubenswrapper[4813]: I1206 16:01:29.984085 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfsj2\" (UniqueName: \"kubernetes.io/projected/1f91a370-ae93-4992-95f9-ebb72180adec-kube-api-access-wfsj2\") pod \"1f91a370-ae93-4992-95f9-ebb72180adec\" (UID: \"1f91a370-ae93-4992-95f9-ebb72180adec\") " Dec 06 16:01:29 crc kubenswrapper[4813]: I1206 16:01:29.984818 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f91a370-ae93-4992-95f9-ebb72180adec-utilities" (OuterVolumeSpecName: "utilities") pod "1f91a370-ae93-4992-95f9-ebb72180adec" (UID: "1f91a370-ae93-4992-95f9-ebb72180adec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:01:30 crc kubenswrapper[4813]: I1206 16:01:30.008208 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f91a370-ae93-4992-95f9-ebb72180adec-kube-api-access-wfsj2" (OuterVolumeSpecName: "kube-api-access-wfsj2") pod "1f91a370-ae93-4992-95f9-ebb72180adec" (UID: "1f91a370-ae93-4992-95f9-ebb72180adec"). InnerVolumeSpecName "kube-api-access-wfsj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:01:30 crc kubenswrapper[4813]: I1206 16:01:30.011510 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f91a370-ae93-4992-95f9-ebb72180adec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1f91a370-ae93-4992-95f9-ebb72180adec" (UID: "1f91a370-ae93-4992-95f9-ebb72180adec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:01:30 crc kubenswrapper[4813]: I1206 16:01:30.086426 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f91a370-ae93-4992-95f9-ebb72180adec-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:30 crc kubenswrapper[4813]: I1206 16:01:30.086457 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f91a370-ae93-4992-95f9-ebb72180adec-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:30 crc kubenswrapper[4813]: I1206 16:01:30.086467 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfsj2\" (UniqueName: \"kubernetes.io/projected/1f91a370-ae93-4992-95f9-ebb72180adec-kube-api-access-wfsj2\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:30 crc kubenswrapper[4813]: I1206 16:01:30.703937 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-2fqgs" event={"ID":"00d42d29-5c9b-4585-be1b-51077e528837","Type":"ContainerStarted","Data":"74e9fab49965bc596784e66a98c531de916ea2736a90c378139c71d01fab0084"} Dec 06 16:01:30 crc kubenswrapper[4813]: I1206 16:01:30.704059 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ccc8479f9-2fqgs" podUID="00d42d29-5c9b-4585-be1b-51077e528837" containerName="dnsmasq-dns" containerID="cri-o://74e9fab49965bc596784e66a98c531de916ea2736a90c378139c71d01fab0084" gracePeriod=10 Dec 06 16:01:30 crc kubenswrapper[4813]: I1206 16:01:30.704286 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5ccc8479f9-2fqgs" Dec 06 16:01:30 crc kubenswrapper[4813]: I1206 16:01:30.708382 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19","Type":"ContainerStarted","Data":"d2852b625a8e7a7b46550b6c2cafd975948130b8c072519d3613885a39ad6c8f"} Dec 06 16:01:30 crc kubenswrapper[4813]: I1206 16:01:30.709701 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420","Type":"ContainerStarted","Data":"d316937a36e31618182a26c06560b3444a08d86af54e55a60ec314c0fa75c45c"} Dec 06 16:01:30 crc kubenswrapper[4813]: I1206 16:01:30.718299 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"75d94fff-d912-4493-b9b9-c458138a7ccf","Type":"ContainerStarted","Data":"c451702f017bff21458ca0ed49a018cc100f3064c5aa5245f97a50ffc3ca9f0e"} Dec 06 16:01:30 crc kubenswrapper[4813]: I1206 16:01:30.719642 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c0a272bd-6444-4c90-acb0-fb653bf3b180","Type":"ContainerStarted","Data":"f0cd9d82918ae2d166e4839ad3cb6421f8554f1f5ee67181f8b7c20095cf4683"} Dec 06 16:01:30 crc kubenswrapper[4813]: I1206 16:01:30.721032 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mk4nr" event={"ID":"6b7cce06-d3db-47cd-a542-a77bfdc42f82","Type":"ContainerStarted","Data":"68d2c6f2c7b4a8357676a71b26ddb05be90e7ffb5f66cf206f372405a522a8d6"} Dec 06 16:01:30 crc kubenswrapper[4813]: I1206 16:01:30.723058 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-bg6cb" event={"ID":"ba6825c1-2137-46d6-a695-6bdd53a13ec4","Type":"ContainerStarted","Data":"c8adbc517c1a8bc9747fd9fd619e4586b4c7926b9d992ef5fb58da7589f771f0"} Dec 06 16:01:30 crc kubenswrapper[4813]: I1206 16:01:30.723630 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-bg6cb" Dec 06 16:01:30 crc kubenswrapper[4813]: I1206 16:01:30.725088 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c","Type":"ContainerStarted","Data":"871393b3fbd041093e2cb1109dad615d39267c2204109c066180823a4c9b79b9"} Dec 06 16:01:30 crc kubenswrapper[4813]: I1206 16:01:30.725677 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 06 16:01:30 crc kubenswrapper[4813]: I1206 16:01:30.727749 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lwvb" event={"ID":"1f91a370-ae93-4992-95f9-ebb72180adec","Type":"ContainerDied","Data":"27895c62025fb9cb7e34371bb4209e4a4b34e5307c101887e00d46a6ec0ea7d8"} Dec 06 16:01:30 crc kubenswrapper[4813]: I1206 16:01:30.727871 4813 scope.go:117] "RemoveContainer" containerID="3c58fba782bd161198ed1d4703b75ac9847279c921a8d696fdca53e276030fa7" Dec 06 16:01:30 crc kubenswrapper[4813]: I1206 16:01:30.728046 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4lwvb" Dec 06 16:01:30 crc kubenswrapper[4813]: I1206 16:01:30.732470 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"bf7377f8-ef23-484a-b66b-90db1cd5b7cd","Type":"ContainerStarted","Data":"11be9a0b56dc4b25d65c390db63af9da4adacc88bca7a4527904c46753dc5249"} Dec 06 16:01:30 crc kubenswrapper[4813]: I1206 16:01:30.750392 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c9961e9d-0fa2-4619-9e19-9b2dfd4af40c","Type":"ContainerStarted","Data":"3f45f463a3c6b69a17b6668feff45caeb8fe7254fc4c2843ae755224b30b8589"} Dec 06 16:01:30 crc kubenswrapper[4813]: I1206 16:01:30.751412 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 06 16:01:30 crc kubenswrapper[4813]: I1206 16:01:30.770284 4813 scope.go:117] "RemoveContainer" containerID="15569de231f50cea091f44a21f523150826f61f883ffb6bacb50bb429212b636" Dec 06 16:01:30 crc kubenswrapper[4813]: I1206 16:01:30.780729 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ccc8479f9-2fqgs" podStartSLOduration=15.580197038 podStartE2EDuration="32.780715948s" podCreationTimestamp="2025-12-06 16:00:58 +0000 UTC" firstStartedPulling="2025-12-06 16:00:59.754248801 +0000 UTC m=+899.645128377" lastFinishedPulling="2025-12-06 16:01:16.954767711 +0000 UTC m=+916.845647287" observedRunningTime="2025-12-06 16:01:30.745696691 +0000 UTC m=+930.636576287" watchObservedRunningTime="2025-12-06 16:01:30.780715948 +0000 UTC m=+930.671595524" Dec 06 16:01:30 crc kubenswrapper[4813]: I1206 16:01:30.802393 4813 scope.go:117] "RemoveContainer" containerID="f2897eba0d29b862597ab3dac690c7afc0393fd65c5f8ac802687c8113c7c543" Dec 06 16:01:30 crc kubenswrapper[4813]: I1206 16:01:30.905378 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=18.174261006 podStartE2EDuration="27.905359216s" podCreationTimestamp="2025-12-06 16:01:03 +0000 UTC" firstStartedPulling="2025-12-06 16:01:17.589109251 +0000 UTC m=+917.479988817" lastFinishedPulling="2025-12-06 16:01:27.320207451 +0000 UTC m=+927.211087027" observedRunningTime="2025-12-06 16:01:30.904687418 +0000 UTC m=+930.795566994" watchObservedRunningTime="2025-12-06 16:01:30.905359216 +0000 UTC m=+930.796238782" Dec 06 16:01:30 crc kubenswrapper[4813]: I1206 16:01:30.929891 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-bg6cb" podStartSLOduration=10.700327203 podStartE2EDuration="21.929872945s" podCreationTimestamp="2025-12-06 16:01:09 +0000 UTC" firstStartedPulling="2025-12-06 16:01:18.15839787 +0000 UTC m=+918.049277446" lastFinishedPulling="2025-12-06 16:01:29.387943602 +0000 UTC m=+929.278823188" observedRunningTime="2025-12-06 16:01:30.928787466 +0000 UTC m=+930.819667042" watchObservedRunningTime="2025-12-06 16:01:30.929872945 +0000 UTC m=+930.820752521" Dec 06 16:01:31 crc kubenswrapper[4813]: I1206 16:01:31.020090 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=14.359290226 podStartE2EDuration="26.020069422s" podCreationTimestamp="2025-12-06 16:01:05 +0000 UTC" firstStartedPulling="2025-12-06 16:01:17.832948926 +0000 UTC m=+917.723828502" lastFinishedPulling="2025-12-06 16:01:29.493728122 +0000 UTC m=+929.384607698" observedRunningTime="2025-12-06 16:01:30.991515756 +0000 UTC m=+930.882395332" watchObservedRunningTime="2025-12-06 16:01:31.020069422 +0000 UTC m=+930.910949008" Dec 06 16:01:31 crc kubenswrapper[4813]: I1206 16:01:31.028917 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4lwvb"] Dec 06 16:01:31 crc kubenswrapper[4813]: I1206 16:01:31.045632 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4lwvb"] Dec 06 16:01:31 crc kubenswrapper[4813]: I1206 16:01:31.155355 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-2fqgs" Dec 06 16:01:31 crc kubenswrapper[4813]: I1206 16:01:31.311669 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00d42d29-5c9b-4585-be1b-51077e528837-config\") pod \"00d42d29-5c9b-4585-be1b-51077e528837\" (UID: \"00d42d29-5c9b-4585-be1b-51077e528837\") " Dec 06 16:01:31 crc kubenswrapper[4813]: I1206 16:01:31.311814 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggxdn\" (UniqueName: \"kubernetes.io/projected/00d42d29-5c9b-4585-be1b-51077e528837-kube-api-access-ggxdn\") pod \"00d42d29-5c9b-4585-be1b-51077e528837\" (UID: \"00d42d29-5c9b-4585-be1b-51077e528837\") " Dec 06 16:01:31 crc kubenswrapper[4813]: I1206 16:01:31.311856 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00d42d29-5c9b-4585-be1b-51077e528837-dns-svc\") pod \"00d42d29-5c9b-4585-be1b-51077e528837\" (UID: \"00d42d29-5c9b-4585-be1b-51077e528837\") " Dec 06 16:01:31 crc kubenswrapper[4813]: I1206 16:01:31.318986 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00d42d29-5c9b-4585-be1b-51077e528837-kube-api-access-ggxdn" (OuterVolumeSpecName: "kube-api-access-ggxdn") pod "00d42d29-5c9b-4585-be1b-51077e528837" (UID: "00d42d29-5c9b-4585-be1b-51077e528837"). InnerVolumeSpecName "kube-api-access-ggxdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:01:31 crc kubenswrapper[4813]: I1206 16:01:31.364693 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00d42d29-5c9b-4585-be1b-51077e528837-config" (OuterVolumeSpecName: "config") pod "00d42d29-5c9b-4585-be1b-51077e528837" (UID: "00d42d29-5c9b-4585-be1b-51077e528837"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:01:31 crc kubenswrapper[4813]: I1206 16:01:31.365972 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00d42d29-5c9b-4585-be1b-51077e528837-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "00d42d29-5c9b-4585-be1b-51077e528837" (UID: "00d42d29-5c9b-4585-be1b-51077e528837"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:01:31 crc kubenswrapper[4813]: I1206 16:01:31.414425 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggxdn\" (UniqueName: \"kubernetes.io/projected/00d42d29-5c9b-4585-be1b-51077e528837-kube-api-access-ggxdn\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:31 crc kubenswrapper[4813]: I1206 16:01:31.414453 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00d42d29-5c9b-4585-be1b-51077e528837-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:31 crc kubenswrapper[4813]: I1206 16:01:31.414463 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00d42d29-5c9b-4585-be1b-51077e528837-config\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:31 crc kubenswrapper[4813]: I1206 16:01:31.761494 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2172dfb5-c0a6-4525-a326-dd9736eb5233","Type":"ContainerStarted","Data":"2a1916b4fe499237d82c88926954303ec3c62edb0721baeedf52bceebb68441b"} Dec 06 16:01:31 crc kubenswrapper[4813]: I1206 16:01:31.764792 4813 generic.go:334] "Generic (PLEG): container finished" podID="6b7cce06-d3db-47cd-a542-a77bfdc42f82" containerID="68d2c6f2c7b4a8357676a71b26ddb05be90e7ffb5f66cf206f372405a522a8d6" exitCode=0 Dec 06 16:01:31 crc kubenswrapper[4813]: I1206 16:01:31.764853 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mk4nr" event={"ID":"6b7cce06-d3db-47cd-a542-a77bfdc42f82","Type":"ContainerDied","Data":"68d2c6f2c7b4a8357676a71b26ddb05be90e7ffb5f66cf206f372405a522a8d6"} Dec 06 16:01:31 crc kubenswrapper[4813]: I1206 16:01:31.770532 4813 generic.go:334] "Generic (PLEG): container finished" podID="00d42d29-5c9b-4585-be1b-51077e528837" containerID="74e9fab49965bc596784e66a98c531de916ea2736a90c378139c71d01fab0084" exitCode=0 Dec 06 16:01:31 crc kubenswrapper[4813]: I1206 16:01:31.770680 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-2fqgs" Dec 06 16:01:31 crc kubenswrapper[4813]: I1206 16:01:31.770785 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-2fqgs" event={"ID":"00d42d29-5c9b-4585-be1b-51077e528837","Type":"ContainerDied","Data":"74e9fab49965bc596784e66a98c531de916ea2736a90c378139c71d01fab0084"} Dec 06 16:01:31 crc kubenswrapper[4813]: I1206 16:01:31.770849 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-2fqgs" event={"ID":"00d42d29-5c9b-4585-be1b-51077e528837","Type":"ContainerDied","Data":"495e4ccbc38c31136a94ca11755afb0728a50af7fc99d86c23845fc9cf0b6dfd"} Dec 06 16:01:31 crc kubenswrapper[4813]: I1206 16:01:31.770884 4813 scope.go:117] "RemoveContainer" containerID="74e9fab49965bc596784e66a98c531de916ea2736a90c378139c71d01fab0084" Dec 06 16:01:31 crc kubenswrapper[4813]: I1206 16:01:31.802389 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-2fqgs"] Dec 06 16:01:31 crc kubenswrapper[4813]: I1206 16:01:31.806501 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-2fqgs"] Dec 06 16:01:31 crc kubenswrapper[4813]: I1206 16:01:31.821549 4813 scope.go:117] "RemoveContainer" containerID="92ab4225b77064b8f8a05203d2ce0a3065fdb43b3f7cf55fb3addb1b45a5805e" Dec 06 16:01:31 crc kubenswrapper[4813]: I1206 16:01:31.868535 4813 scope.go:117] "RemoveContainer" containerID="74e9fab49965bc596784e66a98c531de916ea2736a90c378139c71d01fab0084" Dec 06 16:01:31 crc kubenswrapper[4813]: E1206 16:01:31.869663 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74e9fab49965bc596784e66a98c531de916ea2736a90c378139c71d01fab0084\": container with ID starting with 74e9fab49965bc596784e66a98c531de916ea2736a90c378139c71d01fab0084 not found: ID does not exist" containerID="74e9fab49965bc596784e66a98c531de916ea2736a90c378139c71d01fab0084" Dec 06 16:01:31 crc kubenswrapper[4813]: I1206 16:01:31.869693 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74e9fab49965bc596784e66a98c531de916ea2736a90c378139c71d01fab0084"} err="failed to get container status \"74e9fab49965bc596784e66a98c531de916ea2736a90c378139c71d01fab0084\": rpc error: code = NotFound desc = could not find container \"74e9fab49965bc596784e66a98c531de916ea2736a90c378139c71d01fab0084\": container with ID starting with 74e9fab49965bc596784e66a98c531de916ea2736a90c378139c71d01fab0084 not found: ID does not exist" Dec 06 16:01:31 crc kubenswrapper[4813]: I1206 16:01:31.869716 4813 scope.go:117] "RemoveContainer" containerID="92ab4225b77064b8f8a05203d2ce0a3065fdb43b3f7cf55fb3addb1b45a5805e" Dec 06 16:01:31 crc kubenswrapper[4813]: E1206 16:01:31.877415 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92ab4225b77064b8f8a05203d2ce0a3065fdb43b3f7cf55fb3addb1b45a5805e\": container with ID starting with 92ab4225b77064b8f8a05203d2ce0a3065fdb43b3f7cf55fb3addb1b45a5805e not found: ID does not exist" containerID="92ab4225b77064b8f8a05203d2ce0a3065fdb43b3f7cf55fb3addb1b45a5805e" Dec 06 16:01:31 crc kubenswrapper[4813]: I1206 16:01:31.877467 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92ab4225b77064b8f8a05203d2ce0a3065fdb43b3f7cf55fb3addb1b45a5805e"} err="failed to get container status \"92ab4225b77064b8f8a05203d2ce0a3065fdb43b3f7cf55fb3addb1b45a5805e\": rpc error: code = NotFound desc = could not find container \"92ab4225b77064b8f8a05203d2ce0a3065fdb43b3f7cf55fb3addb1b45a5805e\": container with ID starting with 92ab4225b77064b8f8a05203d2ce0a3065fdb43b3f7cf55fb3addb1b45a5805e not found: ID does not exist" Dec 06 16:01:32 crc kubenswrapper[4813]: I1206 16:01:32.497054 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00d42d29-5c9b-4585-be1b-51077e528837" path="/var/lib/kubelet/pods/00d42d29-5c9b-4585-be1b-51077e528837/volumes" Dec 06 16:01:32 crc kubenswrapper[4813]: I1206 16:01:32.499312 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f91a370-ae93-4992-95f9-ebb72180adec" path="/var/lib/kubelet/pods/1f91a370-ae93-4992-95f9-ebb72180adec/volumes" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.359592 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-9bcrr"] Dec 06 16:01:33 crc kubenswrapper[4813]: E1206 16:01:33.359857 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f91a370-ae93-4992-95f9-ebb72180adec" containerName="registry-server" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.359870 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f91a370-ae93-4992-95f9-ebb72180adec" containerName="registry-server" Dec 06 16:01:33 crc kubenswrapper[4813]: E1206 16:01:33.359882 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9cd8622-0e78-400c-9de3-213fe41d1c1a" containerName="extract-utilities" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.359889 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9cd8622-0e78-400c-9de3-213fe41d1c1a" containerName="extract-utilities" Dec 06 16:01:33 crc kubenswrapper[4813]: E1206 16:01:33.359905 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00d42d29-5c9b-4585-be1b-51077e528837" containerName="init" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.359912 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="00d42d29-5c9b-4585-be1b-51077e528837" containerName="init" Dec 06 16:01:33 crc kubenswrapper[4813]: E1206 16:01:33.359924 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9cd8622-0e78-400c-9de3-213fe41d1c1a" containerName="extract-content" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.359929 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9cd8622-0e78-400c-9de3-213fe41d1c1a" containerName="extract-content" Dec 06 16:01:33 crc kubenswrapper[4813]: E1206 16:01:33.359939 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00d42d29-5c9b-4585-be1b-51077e528837" containerName="dnsmasq-dns" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.359945 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="00d42d29-5c9b-4585-be1b-51077e528837" containerName="dnsmasq-dns" Dec 06 16:01:33 crc kubenswrapper[4813]: E1206 16:01:33.359961 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f91a370-ae93-4992-95f9-ebb72180adec" containerName="extract-content" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.359966 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f91a370-ae93-4992-95f9-ebb72180adec" containerName="extract-content" Dec 06 16:01:33 crc kubenswrapper[4813]: E1206 16:01:33.359978 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9cd8622-0e78-400c-9de3-213fe41d1c1a" containerName="registry-server" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.359983 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9cd8622-0e78-400c-9de3-213fe41d1c1a" containerName="registry-server" Dec 06 16:01:33 crc kubenswrapper[4813]: E1206 16:01:33.359994 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f91a370-ae93-4992-95f9-ebb72180adec" containerName="extract-utilities" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.360000 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f91a370-ae93-4992-95f9-ebb72180adec" containerName="extract-utilities" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.360139 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9cd8622-0e78-400c-9de3-213fe41d1c1a" containerName="registry-server" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.360151 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f91a370-ae93-4992-95f9-ebb72180adec" containerName="registry-server" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.360162 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="00d42d29-5c9b-4585-be1b-51077e528837" containerName="dnsmasq-dns" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.360668 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-9bcrr" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.365698 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.385745 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-9bcrr"] Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.446983 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/da431468-e795-477e-97cd-b1796a76a117-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-9bcrr\" (UID: \"da431468-e795-477e-97cd-b1796a76a117\") " pod="openstack/ovn-controller-metrics-9bcrr" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.447032 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vtsm\" (UniqueName: \"kubernetes.io/projected/da431468-e795-477e-97cd-b1796a76a117-kube-api-access-5vtsm\") pod \"ovn-controller-metrics-9bcrr\" (UID: \"da431468-e795-477e-97cd-b1796a76a117\") " pod="openstack/ovn-controller-metrics-9bcrr" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.447063 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/da431468-e795-477e-97cd-b1796a76a117-ovs-rundir\") pod \"ovn-controller-metrics-9bcrr\" (UID: \"da431468-e795-477e-97cd-b1796a76a117\") " pod="openstack/ovn-controller-metrics-9bcrr" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.447118 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/da431468-e795-477e-97cd-b1796a76a117-ovn-rundir\") pod \"ovn-controller-metrics-9bcrr\" (UID: \"da431468-e795-477e-97cd-b1796a76a117\") " pod="openstack/ovn-controller-metrics-9bcrr" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.447133 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da431468-e795-477e-97cd-b1796a76a117-config\") pod \"ovn-controller-metrics-9bcrr\" (UID: \"da431468-e795-477e-97cd-b1796a76a117\") " pod="openstack/ovn-controller-metrics-9bcrr" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.447166 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da431468-e795-477e-97cd-b1796a76a117-combined-ca-bundle\") pod \"ovn-controller-metrics-9bcrr\" (UID: \"da431468-e795-477e-97cd-b1796a76a117\") " pod="openstack/ovn-controller-metrics-9bcrr" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.549123 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da431468-e795-477e-97cd-b1796a76a117-combined-ca-bundle\") pod \"ovn-controller-metrics-9bcrr\" (UID: \"da431468-e795-477e-97cd-b1796a76a117\") " pod="openstack/ovn-controller-metrics-9bcrr" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.549391 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/da431468-e795-477e-97cd-b1796a76a117-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-9bcrr\" (UID: \"da431468-e795-477e-97cd-b1796a76a117\") " pod="openstack/ovn-controller-metrics-9bcrr" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.549441 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vtsm\" (UniqueName: \"kubernetes.io/projected/da431468-e795-477e-97cd-b1796a76a117-kube-api-access-5vtsm\") pod \"ovn-controller-metrics-9bcrr\" (UID: \"da431468-e795-477e-97cd-b1796a76a117\") " pod="openstack/ovn-controller-metrics-9bcrr" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.549521 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/da431468-e795-477e-97cd-b1796a76a117-ovs-rundir\") pod \"ovn-controller-metrics-9bcrr\" (UID: \"da431468-e795-477e-97cd-b1796a76a117\") " pod="openstack/ovn-controller-metrics-9bcrr" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.549680 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/da431468-e795-477e-97cd-b1796a76a117-ovn-rundir\") pod \"ovn-controller-metrics-9bcrr\" (UID: \"da431468-e795-477e-97cd-b1796a76a117\") " pod="openstack/ovn-controller-metrics-9bcrr" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.549706 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da431468-e795-477e-97cd-b1796a76a117-config\") pod \"ovn-controller-metrics-9bcrr\" (UID: \"da431468-e795-477e-97cd-b1796a76a117\") " pod="openstack/ovn-controller-metrics-9bcrr" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.550523 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/da431468-e795-477e-97cd-b1796a76a117-ovs-rundir\") pod \"ovn-controller-metrics-9bcrr\" (UID: \"da431468-e795-477e-97cd-b1796a76a117\") " pod="openstack/ovn-controller-metrics-9bcrr" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.550721 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/da431468-e795-477e-97cd-b1796a76a117-ovn-rundir\") pod \"ovn-controller-metrics-9bcrr\" (UID: \"da431468-e795-477e-97cd-b1796a76a117\") " pod="openstack/ovn-controller-metrics-9bcrr" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.551219 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da431468-e795-477e-97cd-b1796a76a117-config\") pod \"ovn-controller-metrics-9bcrr\" (UID: \"da431468-e795-477e-97cd-b1796a76a117\") " pod="openstack/ovn-controller-metrics-9bcrr" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.559905 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/da431468-e795-477e-97cd-b1796a76a117-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-9bcrr\" (UID: \"da431468-e795-477e-97cd-b1796a76a117\") " pod="openstack/ovn-controller-metrics-9bcrr" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.560629 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da431468-e795-477e-97cd-b1796a76a117-combined-ca-bundle\") pod \"ovn-controller-metrics-9bcrr\" (UID: \"da431468-e795-477e-97cd-b1796a76a117\") " pod="openstack/ovn-controller-metrics-9bcrr" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.573709 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vtsm\" (UniqueName: \"kubernetes.io/projected/da431468-e795-477e-97cd-b1796a76a117-kube-api-access-5vtsm\") pod \"ovn-controller-metrics-9bcrr\" (UID: \"da431468-e795-477e-97cd-b1796a76a117\") " pod="openstack/ovn-controller-metrics-9bcrr" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.580197 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-d888c"] Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.581354 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-d888c" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.594452 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.608203 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-d888c"] Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.650282 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b64aae5-1105-4826-a279-b1fb35665793-config\") pod \"dnsmasq-dns-5bf47b49b7-d888c\" (UID: \"0b64aae5-1105-4826-a279-b1fb35665793\") " pod="openstack/dnsmasq-dns-5bf47b49b7-d888c" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.650358 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b64aae5-1105-4826-a279-b1fb35665793-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-d888c\" (UID: \"0b64aae5-1105-4826-a279-b1fb35665793\") " pod="openstack/dnsmasq-dns-5bf47b49b7-d888c" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.650386 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xz2g\" (UniqueName: \"kubernetes.io/projected/0b64aae5-1105-4826-a279-b1fb35665793-kube-api-access-8xz2g\") pod \"dnsmasq-dns-5bf47b49b7-d888c\" (UID: \"0b64aae5-1105-4826-a279-b1fb35665793\") " pod="openstack/dnsmasq-dns-5bf47b49b7-d888c" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.650411 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b64aae5-1105-4826-a279-b1fb35665793-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-d888c\" (UID: \"0b64aae5-1105-4826-a279-b1fb35665793\") " pod="openstack/dnsmasq-dns-5bf47b49b7-d888c" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.675742 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-9bcrr" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.752644 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b64aae5-1105-4826-a279-b1fb35665793-config\") pod \"dnsmasq-dns-5bf47b49b7-d888c\" (UID: \"0b64aae5-1105-4826-a279-b1fb35665793\") " pod="openstack/dnsmasq-dns-5bf47b49b7-d888c" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.752710 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b64aae5-1105-4826-a279-b1fb35665793-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-d888c\" (UID: \"0b64aae5-1105-4826-a279-b1fb35665793\") " pod="openstack/dnsmasq-dns-5bf47b49b7-d888c" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.752734 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xz2g\" (UniqueName: \"kubernetes.io/projected/0b64aae5-1105-4826-a279-b1fb35665793-kube-api-access-8xz2g\") pod \"dnsmasq-dns-5bf47b49b7-d888c\" (UID: \"0b64aae5-1105-4826-a279-b1fb35665793\") " pod="openstack/dnsmasq-dns-5bf47b49b7-d888c" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.752754 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b64aae5-1105-4826-a279-b1fb35665793-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-d888c\" (UID: \"0b64aae5-1105-4826-a279-b1fb35665793\") " pod="openstack/dnsmasq-dns-5bf47b49b7-d888c" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.753594 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b64aae5-1105-4826-a279-b1fb35665793-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-d888c\" (UID: \"0b64aae5-1105-4826-a279-b1fb35665793\") " pod="openstack/dnsmasq-dns-5bf47b49b7-d888c" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.754559 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b64aae5-1105-4826-a279-b1fb35665793-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-d888c\" (UID: \"0b64aae5-1105-4826-a279-b1fb35665793\") " pod="openstack/dnsmasq-dns-5bf47b49b7-d888c" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.755220 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b64aae5-1105-4826-a279-b1fb35665793-config\") pod \"dnsmasq-dns-5bf47b49b7-d888c\" (UID: \"0b64aae5-1105-4826-a279-b1fb35665793\") " pod="openstack/dnsmasq-dns-5bf47b49b7-d888c" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.786642 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xz2g\" (UniqueName: \"kubernetes.io/projected/0b64aae5-1105-4826-a279-b1fb35665793-kube-api-access-8xz2g\") pod \"dnsmasq-dns-5bf47b49b7-d888c\" (UID: \"0b64aae5-1105-4826-a279-b1fb35665793\") " pod="openstack/dnsmasq-dns-5bf47b49b7-d888c" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.869599 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-d888c"] Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.870556 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-d888c" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.915741 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-kckf9"] Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.917025 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-kckf9" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.919190 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 06 16:01:33 crc kubenswrapper[4813]: I1206 16:01:33.950749 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-kckf9"] Dec 06 16:01:34 crc kubenswrapper[4813]: I1206 16:01:34.057209 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lqhn\" (UniqueName: \"kubernetes.io/projected/0497b076-9f37-46bb-a5dd-370f89e97b15-kube-api-access-7lqhn\") pod \"dnsmasq-dns-8554648995-kckf9\" (UID: \"0497b076-9f37-46bb-a5dd-370f89e97b15\") " pod="openstack/dnsmasq-dns-8554648995-kckf9" Dec 06 16:01:34 crc kubenswrapper[4813]: I1206 16:01:34.057257 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0497b076-9f37-46bb-a5dd-370f89e97b15-dns-svc\") pod \"dnsmasq-dns-8554648995-kckf9\" (UID: \"0497b076-9f37-46bb-a5dd-370f89e97b15\") " pod="openstack/dnsmasq-dns-8554648995-kckf9" Dec 06 16:01:34 crc kubenswrapper[4813]: I1206 16:01:34.057331 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0497b076-9f37-46bb-a5dd-370f89e97b15-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-kckf9\" (UID: \"0497b076-9f37-46bb-a5dd-370f89e97b15\") " pod="openstack/dnsmasq-dns-8554648995-kckf9" Dec 06 16:01:34 crc kubenswrapper[4813]: I1206 16:01:34.057372 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0497b076-9f37-46bb-a5dd-370f89e97b15-config\") pod \"dnsmasq-dns-8554648995-kckf9\" (UID: \"0497b076-9f37-46bb-a5dd-370f89e97b15\") " pod="openstack/dnsmasq-dns-8554648995-kckf9" Dec 06 16:01:34 crc kubenswrapper[4813]: I1206 16:01:34.057392 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0497b076-9f37-46bb-a5dd-370f89e97b15-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-kckf9\" (UID: \"0497b076-9f37-46bb-a5dd-370f89e97b15\") " pod="openstack/dnsmasq-dns-8554648995-kckf9" Dec 06 16:01:34 crc kubenswrapper[4813]: I1206 16:01:34.137815 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-9bcrr"] Dec 06 16:01:34 crc kubenswrapper[4813]: I1206 16:01:34.158282 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0497b076-9f37-46bb-a5dd-370f89e97b15-config\") pod \"dnsmasq-dns-8554648995-kckf9\" (UID: \"0497b076-9f37-46bb-a5dd-370f89e97b15\") " pod="openstack/dnsmasq-dns-8554648995-kckf9" Dec 06 16:01:34 crc kubenswrapper[4813]: I1206 16:01:34.158537 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0497b076-9f37-46bb-a5dd-370f89e97b15-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-kckf9\" (UID: \"0497b076-9f37-46bb-a5dd-370f89e97b15\") " pod="openstack/dnsmasq-dns-8554648995-kckf9" Dec 06 16:01:34 crc kubenswrapper[4813]: I1206 16:01:34.158590 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lqhn\" (UniqueName: \"kubernetes.io/projected/0497b076-9f37-46bb-a5dd-370f89e97b15-kube-api-access-7lqhn\") pod \"dnsmasq-dns-8554648995-kckf9\" (UID: \"0497b076-9f37-46bb-a5dd-370f89e97b15\") " pod="openstack/dnsmasq-dns-8554648995-kckf9" Dec 06 16:01:34 crc kubenswrapper[4813]: I1206 16:01:34.158612 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0497b076-9f37-46bb-a5dd-370f89e97b15-dns-svc\") pod \"dnsmasq-dns-8554648995-kckf9\" (UID: \"0497b076-9f37-46bb-a5dd-370f89e97b15\") " pod="openstack/dnsmasq-dns-8554648995-kckf9" Dec 06 16:01:34 crc kubenswrapper[4813]: I1206 16:01:34.158660 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0497b076-9f37-46bb-a5dd-370f89e97b15-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-kckf9\" (UID: \"0497b076-9f37-46bb-a5dd-370f89e97b15\") " pod="openstack/dnsmasq-dns-8554648995-kckf9" Dec 06 16:01:34 crc kubenswrapper[4813]: I1206 16:01:34.159561 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0497b076-9f37-46bb-a5dd-370f89e97b15-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-kckf9\" (UID: \"0497b076-9f37-46bb-a5dd-370f89e97b15\") " pod="openstack/dnsmasq-dns-8554648995-kckf9" Dec 06 16:01:34 crc kubenswrapper[4813]: I1206 16:01:34.160011 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0497b076-9f37-46bb-a5dd-370f89e97b15-config\") pod \"dnsmasq-dns-8554648995-kckf9\" (UID: \"0497b076-9f37-46bb-a5dd-370f89e97b15\") " pod="openstack/dnsmasq-dns-8554648995-kckf9" Dec 06 16:01:34 crc kubenswrapper[4813]: I1206 16:01:34.160067 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0497b076-9f37-46bb-a5dd-370f89e97b15-dns-svc\") pod \"dnsmasq-dns-8554648995-kckf9\" (UID: \"0497b076-9f37-46bb-a5dd-370f89e97b15\") " pod="openstack/dnsmasq-dns-8554648995-kckf9" Dec 06 16:01:34 crc kubenswrapper[4813]: I1206 16:01:34.160510 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0497b076-9f37-46bb-a5dd-370f89e97b15-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-kckf9\" (UID: \"0497b076-9f37-46bb-a5dd-370f89e97b15\") " pod="openstack/dnsmasq-dns-8554648995-kckf9" Dec 06 16:01:34 crc kubenswrapper[4813]: I1206 16:01:34.195717 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lqhn\" (UniqueName: \"kubernetes.io/projected/0497b076-9f37-46bb-a5dd-370f89e97b15-kube-api-access-7lqhn\") pod \"dnsmasq-dns-8554648995-kckf9\" (UID: \"0497b076-9f37-46bb-a5dd-370f89e97b15\") " pod="openstack/dnsmasq-dns-8554648995-kckf9" Dec 06 16:01:34 crc kubenswrapper[4813]: I1206 16:01:34.242399 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-kckf9" Dec 06 16:01:34 crc kubenswrapper[4813]: I1206 16:01:34.499219 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-d888c"] Dec 06 16:01:34 crc kubenswrapper[4813]: W1206 16:01:34.720384 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0497b076_9f37_46bb_a5dd_370f89e97b15.slice/crio-00eb185060e50110c7144725493e065052bffe08d9c3075c9222f141ab00fcff WatchSource:0}: Error finding container 00eb185060e50110c7144725493e065052bffe08d9c3075c9222f141ab00fcff: Status 404 returned error can't find the container with id 00eb185060e50110c7144725493e065052bffe08d9c3075c9222f141ab00fcff Dec 06 16:01:34 crc kubenswrapper[4813]: I1206 16:01:34.722288 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-kckf9"] Dec 06 16:01:34 crc kubenswrapper[4813]: I1206 16:01:34.820451 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-kckf9" event={"ID":"0497b076-9f37-46bb-a5dd-370f89e97b15","Type":"ContainerStarted","Data":"00eb185060e50110c7144725493e065052bffe08d9c3075c9222f141ab00fcff"} Dec 06 16:01:34 crc kubenswrapper[4813]: I1206 16:01:34.822157 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-d888c" event={"ID":"0b64aae5-1105-4826-a279-b1fb35665793","Type":"ContainerStarted","Data":"a47669ef22bf7502aa64faf6e934d564a6ae205cca1b58c5421e2e39bda997cf"} Dec 06 16:01:34 crc kubenswrapper[4813]: I1206 16:01:34.823439 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-9bcrr" event={"ID":"da431468-e795-477e-97cd-b1796a76a117","Type":"ContainerStarted","Data":"a0bd53a3659eb8c0ba2e008dee009959dd91a697a0b08b4f0b24a47c7324ce7d"} Dec 06 16:01:35 crc kubenswrapper[4813]: I1206 16:01:35.698556 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 06 16:01:35 crc kubenswrapper[4813]: I1206 16:01:35.831104 4813 generic.go:334] "Generic (PLEG): container finished" podID="0b64aae5-1105-4826-a279-b1fb35665793" containerID="14f3f9eccd00ab79159dd805cd3bba9e94770e8c5e364aedc0d403e2cdb1cf9a" exitCode=0 Dec 06 16:01:35 crc kubenswrapper[4813]: I1206 16:01:35.831171 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-d888c" event={"ID":"0b64aae5-1105-4826-a279-b1fb35665793","Type":"ContainerDied","Data":"14f3f9eccd00ab79159dd805cd3bba9e94770e8c5e364aedc0d403e2cdb1cf9a"} Dec 06 16:01:35 crc kubenswrapper[4813]: I1206 16:01:35.834664 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mk4nr" event={"ID":"6b7cce06-d3db-47cd-a542-a77bfdc42f82","Type":"ContainerStarted","Data":"1f0957c82cac05c7b81e718e18e9c327ed2066fce44795b409b10b3b3f6ceaf8"} Dec 06 16:01:35 crc kubenswrapper[4813]: I1206 16:01:35.836405 4813 generic.go:334] "Generic (PLEG): container finished" podID="0497b076-9f37-46bb-a5dd-370f89e97b15" containerID="a9551fb18c6dd23ac19ebb75211a695bc01a9c410d898e26b1ea02ad90953574" exitCode=0 Dec 06 16:01:35 crc kubenswrapper[4813]: I1206 16:01:35.836441 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-kckf9" event={"ID":"0497b076-9f37-46bb-a5dd-370f89e97b15","Type":"ContainerDied","Data":"a9551fb18c6dd23ac19ebb75211a695bc01a9c410d898e26b1ea02ad90953574"} Dec 06 16:01:37 crc kubenswrapper[4813]: I1206 16:01:37.593600 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-d888c" Dec 06 16:01:37 crc kubenswrapper[4813]: I1206 16:01:37.645294 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b64aae5-1105-4826-a279-b1fb35665793-config\") pod \"0b64aae5-1105-4826-a279-b1fb35665793\" (UID: \"0b64aae5-1105-4826-a279-b1fb35665793\") " Dec 06 16:01:37 crc kubenswrapper[4813]: I1206 16:01:37.645343 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b64aae5-1105-4826-a279-b1fb35665793-dns-svc\") pod \"0b64aae5-1105-4826-a279-b1fb35665793\" (UID: \"0b64aae5-1105-4826-a279-b1fb35665793\") " Dec 06 16:01:37 crc kubenswrapper[4813]: I1206 16:01:37.645492 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xz2g\" (UniqueName: \"kubernetes.io/projected/0b64aae5-1105-4826-a279-b1fb35665793-kube-api-access-8xz2g\") pod \"0b64aae5-1105-4826-a279-b1fb35665793\" (UID: \"0b64aae5-1105-4826-a279-b1fb35665793\") " Dec 06 16:01:37 crc kubenswrapper[4813]: I1206 16:01:37.645514 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b64aae5-1105-4826-a279-b1fb35665793-ovsdbserver-nb\") pod \"0b64aae5-1105-4826-a279-b1fb35665793\" (UID: \"0b64aae5-1105-4826-a279-b1fb35665793\") " Dec 06 16:01:37 crc kubenswrapper[4813]: I1206 16:01:37.657334 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b64aae5-1105-4826-a279-b1fb35665793-kube-api-access-8xz2g" (OuterVolumeSpecName: "kube-api-access-8xz2g") pod "0b64aae5-1105-4826-a279-b1fb35665793" (UID: "0b64aae5-1105-4826-a279-b1fb35665793"). InnerVolumeSpecName "kube-api-access-8xz2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:01:37 crc kubenswrapper[4813]: I1206 16:01:37.664450 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b64aae5-1105-4826-a279-b1fb35665793-config" (OuterVolumeSpecName: "config") pod "0b64aae5-1105-4826-a279-b1fb35665793" (UID: "0b64aae5-1105-4826-a279-b1fb35665793"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:01:37 crc kubenswrapper[4813]: I1206 16:01:37.666722 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b64aae5-1105-4826-a279-b1fb35665793-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0b64aae5-1105-4826-a279-b1fb35665793" (UID: "0b64aae5-1105-4826-a279-b1fb35665793"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:01:37 crc kubenswrapper[4813]: I1206 16:01:37.680553 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b64aae5-1105-4826-a279-b1fb35665793-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0b64aae5-1105-4826-a279-b1fb35665793" (UID: "0b64aae5-1105-4826-a279-b1fb35665793"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:01:37 crc kubenswrapper[4813]: I1206 16:01:37.746817 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xz2g\" (UniqueName: \"kubernetes.io/projected/0b64aae5-1105-4826-a279-b1fb35665793-kube-api-access-8xz2g\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:37 crc kubenswrapper[4813]: I1206 16:01:37.746849 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b64aae5-1105-4826-a279-b1fb35665793-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:37 crc kubenswrapper[4813]: I1206 16:01:37.746863 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b64aae5-1105-4826-a279-b1fb35665793-config\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:37 crc kubenswrapper[4813]: I1206 16:01:37.746875 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b64aae5-1105-4826-a279-b1fb35665793-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:37 crc kubenswrapper[4813]: I1206 16:01:37.862779 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-d888c" event={"ID":"0b64aae5-1105-4826-a279-b1fb35665793","Type":"ContainerDied","Data":"a47669ef22bf7502aa64faf6e934d564a6ae205cca1b58c5421e2e39bda997cf"} Dec 06 16:01:37 crc kubenswrapper[4813]: I1206 16:01:37.863033 4813 scope.go:117] "RemoveContainer" containerID="14f3f9eccd00ab79159dd805cd3bba9e94770e8c5e364aedc0d403e2cdb1cf9a" Dec 06 16:01:37 crc kubenswrapper[4813]: I1206 16:01:37.862857 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-d888c" Dec 06 16:01:37 crc kubenswrapper[4813]: I1206 16:01:37.873639 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mk4nr" event={"ID":"6b7cce06-d3db-47cd-a542-a77bfdc42f82","Type":"ContainerStarted","Data":"36ab7a186879ee295e26f11e3b92b6aacf03f7204b7adcecb284752d6fce6ecb"} Dec 06 16:01:37 crc kubenswrapper[4813]: I1206 16:01:37.874888 4813 generic.go:334] "Generic (PLEG): container finished" podID="bf7377f8-ef23-484a-b66b-90db1cd5b7cd" containerID="11be9a0b56dc4b25d65c390db63af9da4adacc88bca7a4527904c46753dc5249" exitCode=0 Dec 06 16:01:37 crc kubenswrapper[4813]: I1206 16:01:37.874926 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"bf7377f8-ef23-484a-b66b-90db1cd5b7cd","Type":"ContainerDied","Data":"11be9a0b56dc4b25d65c390db63af9da4adacc88bca7a4527904c46753dc5249"} Dec 06 16:01:37 crc kubenswrapper[4813]: I1206 16:01:37.879436 4813 generic.go:334] "Generic (PLEG): container finished" podID="75d94fff-d912-4493-b9b9-c458138a7ccf" containerID="c451702f017bff21458ca0ed49a018cc100f3064c5aa5245f97a50ffc3ca9f0e" exitCode=0 Dec 06 16:01:37 crc kubenswrapper[4813]: I1206 16:01:37.879477 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"75d94fff-d912-4493-b9b9-c458138a7ccf","Type":"ContainerDied","Data":"c451702f017bff21458ca0ed49a018cc100f3064c5aa5245f97a50ffc3ca9f0e"} Dec 06 16:01:37 crc kubenswrapper[4813]: I1206 16:01:37.988559 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-d888c"] Dec 06 16:01:37 crc kubenswrapper[4813]: I1206 16:01:37.995001 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-d888c"] Dec 06 16:01:38 crc kubenswrapper[4813]: I1206 16:01:38.497597 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b64aae5-1105-4826-a279-b1fb35665793" path="/var/lib/kubelet/pods/0b64aae5-1105-4826-a279-b1fb35665793/volumes" Dec 06 16:01:38 crc kubenswrapper[4813]: I1206 16:01:38.865390 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 06 16:01:40 crc kubenswrapper[4813]: I1206 16:01:40.908101 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-kckf9" event={"ID":"0497b076-9f37-46bb-a5dd-370f89e97b15","Type":"ContainerStarted","Data":"0b85051277ec559995330ca64a4036a37971d42cb977db03710e4f19d50d2401"} Dec 06 16:01:40 crc kubenswrapper[4813]: I1206 16:01:40.908958 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-kckf9" Dec 06 16:01:40 crc kubenswrapper[4813]: I1206 16:01:40.914753 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"75d94fff-d912-4493-b9b9-c458138a7ccf","Type":"ContainerStarted","Data":"c43e3dedb4f86946df97d9a609f38130487071175fb2f9a59d1a57906c70b090"} Dec 06 16:01:40 crc kubenswrapper[4813]: I1206 16:01:40.921049 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"bf7377f8-ef23-484a-b66b-90db1cd5b7cd","Type":"ContainerStarted","Data":"4c963076dc24b28f29541c662937dbb87d5dde68e1c3e950fc6c67bc4e98bc66"} Dec 06 16:01:40 crc kubenswrapper[4813]: I1206 16:01:40.921091 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-mk4nr" Dec 06 16:01:40 crc kubenswrapper[4813]: I1206 16:01:40.921112 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-mk4nr" Dec 06 16:01:40 crc kubenswrapper[4813]: I1206 16:01:40.931327 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-kckf9" podStartSLOduration=7.931307492 podStartE2EDuration="7.931307492s" podCreationTimestamp="2025-12-06 16:01:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:01:40.93010708 +0000 UTC m=+940.820986676" watchObservedRunningTime="2025-12-06 16:01:40.931307492 +0000 UTC m=+940.822187068" Dec 06 16:01:40 crc kubenswrapper[4813]: I1206 16:01:40.964466 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-mk4nr" podStartSLOduration=21.615357461 podStartE2EDuration="31.964440238s" podCreationTimestamp="2025-12-06 16:01:09 +0000 UTC" firstStartedPulling="2025-12-06 16:01:18.361543487 +0000 UTC m=+918.252423063" lastFinishedPulling="2025-12-06 16:01:28.710626254 +0000 UTC m=+928.601505840" observedRunningTime="2025-12-06 16:01:40.954025433 +0000 UTC m=+940.844905029" watchObservedRunningTime="2025-12-06 16:01:40.964440238 +0000 UTC m=+940.855319834" Dec 06 16:01:40 crc kubenswrapper[4813]: I1206 16:01:40.982436 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=29.443086782 podStartE2EDuration="40.982412164s" podCreationTimestamp="2025-12-06 16:01:00 +0000 UTC" firstStartedPulling="2025-12-06 16:01:17.827600694 +0000 UTC m=+917.718480270" lastFinishedPulling="2025-12-06 16:01:29.366926086 +0000 UTC m=+929.257805652" observedRunningTime="2025-12-06 16:01:40.974254738 +0000 UTC m=+940.865134324" watchObservedRunningTime="2025-12-06 16:01:40.982412164 +0000 UTC m=+940.873291780" Dec 06 16:01:41 crc kubenswrapper[4813]: I1206 16:01:41.007917 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=27.902470002 podStartE2EDuration="39.007893929s" podCreationTimestamp="2025-12-06 16:01:02 +0000 UTC" firstStartedPulling="2025-12-06 16:01:17.605211848 +0000 UTC m=+917.496091424" lastFinishedPulling="2025-12-06 16:01:28.710635775 +0000 UTC m=+928.601515351" observedRunningTime="2025-12-06 16:01:40.997471673 +0000 UTC m=+940.888351249" watchObservedRunningTime="2025-12-06 16:01:41.007893929 +0000 UTC m=+940.898773505" Dec 06 16:01:41 crc kubenswrapper[4813]: I1206 16:01:41.933699 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-9bcrr" event={"ID":"da431468-e795-477e-97cd-b1796a76a117","Type":"ContainerStarted","Data":"2229c0f652a269af8c8ec6f2f76554cb0018f0b0203c79d869c2fef320f5e77e"} Dec 06 16:01:41 crc kubenswrapper[4813]: I1206 16:01:41.940280 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420","Type":"ContainerStarted","Data":"2345fbfb241eac7d853b2a3e38f73a089418335721e8cb67be16fe68e189434e"} Dec 06 16:01:41 crc kubenswrapper[4813]: I1206 16:01:41.943987 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c0a272bd-6444-4c90-acb0-fb653bf3b180","Type":"ContainerStarted","Data":"9b8bd23cf8de07aa47b28a122859b94c056ed54aceef50cfa255a2f9b4805599"} Dec 06 16:01:41 crc kubenswrapper[4813]: I1206 16:01:41.966177 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-9bcrr" podStartSLOduration=2.361278929 podStartE2EDuration="8.966146712s" podCreationTimestamp="2025-12-06 16:01:33 +0000 UTC" firstStartedPulling="2025-12-06 16:01:34.139546652 +0000 UTC m=+934.030426228" lastFinishedPulling="2025-12-06 16:01:40.744414415 +0000 UTC m=+940.635294011" observedRunningTime="2025-12-06 16:01:41.960819561 +0000 UTC m=+941.851699147" watchObservedRunningTime="2025-12-06 16:01:41.966146712 +0000 UTC m=+941.857026298" Dec 06 16:01:41 crc kubenswrapper[4813]: I1206 16:01:41.997439 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=8.880684355 podStartE2EDuration="29.997411939s" podCreationTimestamp="2025-12-06 16:01:12 +0000 UTC" firstStartedPulling="2025-12-06 16:01:19.62046882 +0000 UTC m=+919.511348396" lastFinishedPulling="2025-12-06 16:01:40.737196384 +0000 UTC m=+940.628075980" observedRunningTime="2025-12-06 16:01:41.993427564 +0000 UTC m=+941.884307150" watchObservedRunningTime="2025-12-06 16:01:41.997411939 +0000 UTC m=+941.888291555" Dec 06 16:01:42 crc kubenswrapper[4813]: I1206 16:01:42.037517 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=13.509526398 podStartE2EDuration="33.03749487s" podCreationTimestamp="2025-12-06 16:01:09 +0000 UTC" firstStartedPulling="2025-12-06 16:01:21.185054802 +0000 UTC m=+921.075934378" lastFinishedPulling="2025-12-06 16:01:40.713023274 +0000 UTC m=+940.603902850" observedRunningTime="2025-12-06 16:01:42.029973431 +0000 UTC m=+941.920853017" watchObservedRunningTime="2025-12-06 16:01:42.03749487 +0000 UTC m=+941.928374456" Dec 06 16:01:42 crc kubenswrapper[4813]: I1206 16:01:42.070851 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 06 16:01:42 crc kubenswrapper[4813]: I1206 16:01:42.070914 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 06 16:01:43 crc kubenswrapper[4813]: I1206 16:01:43.420668 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:43 crc kubenswrapper[4813]: I1206 16:01:43.421104 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:43 crc kubenswrapper[4813]: I1206 16:01:43.572497 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:43 crc kubenswrapper[4813]: I1206 16:01:43.572558 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:43 crc kubenswrapper[4813]: I1206 16:01:43.620637 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:43 crc kubenswrapper[4813]: I1206 16:01:43.998701 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 06 16:01:44 crc kubenswrapper[4813]: I1206 16:01:44.178434 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:44 crc kubenswrapper[4813]: I1206 16:01:44.223541 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:44 crc kubenswrapper[4813]: I1206 16:01:44.969443 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.015473 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.233733 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 06 16:01:45 crc kubenswrapper[4813]: E1206 16:01:45.234039 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b64aae5-1105-4826-a279-b1fb35665793" containerName="init" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.234054 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b64aae5-1105-4826-a279-b1fb35665793" containerName="init" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.234204 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b64aae5-1105-4826-a279-b1fb35665793" containerName="init" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.234965 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.236836 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.236992 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.237484 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.253906 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-j54gh" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.292225 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.294153 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjlkv\" (UniqueName: \"kubernetes.io/projected/9614929b-2a14-4d49-8009-8f627fb000fd-kube-api-access-vjlkv\") pod \"ovn-northd-0\" (UID: \"9614929b-2a14-4d49-8009-8f627fb000fd\") " pod="openstack/ovn-northd-0" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.294219 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9614929b-2a14-4d49-8009-8f627fb000fd-config\") pod \"ovn-northd-0\" (UID: \"9614929b-2a14-4d49-8009-8f627fb000fd\") " pod="openstack/ovn-northd-0" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.294274 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/9614929b-2a14-4d49-8009-8f627fb000fd-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"9614929b-2a14-4d49-8009-8f627fb000fd\") " pod="openstack/ovn-northd-0" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.294322 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9614929b-2a14-4d49-8009-8f627fb000fd-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"9614929b-2a14-4d49-8009-8f627fb000fd\") " pod="openstack/ovn-northd-0" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.294380 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9614929b-2a14-4d49-8009-8f627fb000fd-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"9614929b-2a14-4d49-8009-8f627fb000fd\") " pod="openstack/ovn-northd-0" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.294492 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9614929b-2a14-4d49-8009-8f627fb000fd-scripts\") pod \"ovn-northd-0\" (UID: \"9614929b-2a14-4d49-8009-8f627fb000fd\") " pod="openstack/ovn-northd-0" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.294629 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9614929b-2a14-4d49-8009-8f627fb000fd-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"9614929b-2a14-4d49-8009-8f627fb000fd\") " pod="openstack/ovn-northd-0" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.395605 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjlkv\" (UniqueName: \"kubernetes.io/projected/9614929b-2a14-4d49-8009-8f627fb000fd-kube-api-access-vjlkv\") pod \"ovn-northd-0\" (UID: \"9614929b-2a14-4d49-8009-8f627fb000fd\") " pod="openstack/ovn-northd-0" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.395662 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9614929b-2a14-4d49-8009-8f627fb000fd-config\") pod \"ovn-northd-0\" (UID: \"9614929b-2a14-4d49-8009-8f627fb000fd\") " pod="openstack/ovn-northd-0" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.395689 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/9614929b-2a14-4d49-8009-8f627fb000fd-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"9614929b-2a14-4d49-8009-8f627fb000fd\") " pod="openstack/ovn-northd-0" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.395722 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9614929b-2a14-4d49-8009-8f627fb000fd-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"9614929b-2a14-4d49-8009-8f627fb000fd\") " pod="openstack/ovn-northd-0" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.395747 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9614929b-2a14-4d49-8009-8f627fb000fd-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"9614929b-2a14-4d49-8009-8f627fb000fd\") " pod="openstack/ovn-northd-0" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.395768 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9614929b-2a14-4d49-8009-8f627fb000fd-scripts\") pod \"ovn-northd-0\" (UID: \"9614929b-2a14-4d49-8009-8f627fb000fd\") " pod="openstack/ovn-northd-0" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.395798 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9614929b-2a14-4d49-8009-8f627fb000fd-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"9614929b-2a14-4d49-8009-8f627fb000fd\") " pod="openstack/ovn-northd-0" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.396617 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9614929b-2a14-4d49-8009-8f627fb000fd-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"9614929b-2a14-4d49-8009-8f627fb000fd\") " pod="openstack/ovn-northd-0" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.397007 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9614929b-2a14-4d49-8009-8f627fb000fd-scripts\") pod \"ovn-northd-0\" (UID: \"9614929b-2a14-4d49-8009-8f627fb000fd\") " pod="openstack/ovn-northd-0" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.397128 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9614929b-2a14-4d49-8009-8f627fb000fd-config\") pod \"ovn-northd-0\" (UID: \"9614929b-2a14-4d49-8009-8f627fb000fd\") " pod="openstack/ovn-northd-0" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.402010 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/9614929b-2a14-4d49-8009-8f627fb000fd-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"9614929b-2a14-4d49-8009-8f627fb000fd\") " pod="openstack/ovn-northd-0" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.402784 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9614929b-2a14-4d49-8009-8f627fb000fd-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"9614929b-2a14-4d49-8009-8f627fb000fd\") " pod="openstack/ovn-northd-0" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.410989 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9614929b-2a14-4d49-8009-8f627fb000fd-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"9614929b-2a14-4d49-8009-8f627fb000fd\") " pod="openstack/ovn-northd-0" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.429869 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjlkv\" (UniqueName: \"kubernetes.io/projected/9614929b-2a14-4d49-8009-8f627fb000fd-kube-api-access-vjlkv\") pod \"ovn-northd-0\" (UID: \"9614929b-2a14-4d49-8009-8f627fb000fd\") " pod="openstack/ovn-northd-0" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.553603 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.717706 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-kckf9"] Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.718137 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-kckf9" podUID="0497b076-9f37-46bb-a5dd-370f89e97b15" containerName="dnsmasq-dns" containerID="cri-o://0b85051277ec559995330ca64a4036a37971d42cb977db03710e4f19d50d2401" gracePeriod=10 Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.719621 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-kckf9" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.805725 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-8l7jf"] Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.806964 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-8l7jf" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.882763 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-8l7jf"] Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.907235 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4400510-3d90-46c6-ba53-d964f310719a-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-8l7jf\" (UID: \"b4400510-3d90-46c6-ba53-d964f310719a\") " pod="openstack/dnsmasq-dns-b8fbc5445-8l7jf" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.907342 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2wzt\" (UniqueName: \"kubernetes.io/projected/b4400510-3d90-46c6-ba53-d964f310719a-kube-api-access-x2wzt\") pod \"dnsmasq-dns-b8fbc5445-8l7jf\" (UID: \"b4400510-3d90-46c6-ba53-d964f310719a\") " pod="openstack/dnsmasq-dns-b8fbc5445-8l7jf" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.907371 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4400510-3d90-46c6-ba53-d964f310719a-config\") pod \"dnsmasq-dns-b8fbc5445-8l7jf\" (UID: \"b4400510-3d90-46c6-ba53-d964f310719a\") " pod="openstack/dnsmasq-dns-b8fbc5445-8l7jf" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.907398 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4400510-3d90-46c6-ba53-d964f310719a-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-8l7jf\" (UID: \"b4400510-3d90-46c6-ba53-d964f310719a\") " pod="openstack/dnsmasq-dns-b8fbc5445-8l7jf" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.907422 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4400510-3d90-46c6-ba53-d964f310719a-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-8l7jf\" (UID: \"b4400510-3d90-46c6-ba53-d964f310719a\") " pod="openstack/dnsmasq-dns-b8fbc5445-8l7jf" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.971373 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.995801 4813 generic.go:334] "Generic (PLEG): container finished" podID="0497b076-9f37-46bb-a5dd-370f89e97b15" containerID="0b85051277ec559995330ca64a4036a37971d42cb977db03710e4f19d50d2401" exitCode=0 Dec 06 16:01:45 crc kubenswrapper[4813]: I1206 16:01:45.996562 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-kckf9" event={"ID":"0497b076-9f37-46bb-a5dd-370f89e97b15","Type":"ContainerDied","Data":"0b85051277ec559995330ca64a4036a37971d42cb977db03710e4f19d50d2401"} Dec 06 16:01:46 crc kubenswrapper[4813]: I1206 16:01:46.014663 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4400510-3d90-46c6-ba53-d964f310719a-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-8l7jf\" (UID: \"b4400510-3d90-46c6-ba53-d964f310719a\") " pod="openstack/dnsmasq-dns-b8fbc5445-8l7jf" Dec 06 16:01:46 crc kubenswrapper[4813]: I1206 16:01:46.014701 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2wzt\" (UniqueName: \"kubernetes.io/projected/b4400510-3d90-46c6-ba53-d964f310719a-kube-api-access-x2wzt\") pod \"dnsmasq-dns-b8fbc5445-8l7jf\" (UID: \"b4400510-3d90-46c6-ba53-d964f310719a\") " pod="openstack/dnsmasq-dns-b8fbc5445-8l7jf" Dec 06 16:01:46 crc kubenswrapper[4813]: I1206 16:01:46.014732 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4400510-3d90-46c6-ba53-d964f310719a-config\") pod \"dnsmasq-dns-b8fbc5445-8l7jf\" (UID: \"b4400510-3d90-46c6-ba53-d964f310719a\") " pod="openstack/dnsmasq-dns-b8fbc5445-8l7jf" Dec 06 16:01:46 crc kubenswrapper[4813]: I1206 16:01:46.014761 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4400510-3d90-46c6-ba53-d964f310719a-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-8l7jf\" (UID: \"b4400510-3d90-46c6-ba53-d964f310719a\") " pod="openstack/dnsmasq-dns-b8fbc5445-8l7jf" Dec 06 16:01:46 crc kubenswrapper[4813]: I1206 16:01:46.014777 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4400510-3d90-46c6-ba53-d964f310719a-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-8l7jf\" (UID: \"b4400510-3d90-46c6-ba53-d964f310719a\") " pod="openstack/dnsmasq-dns-b8fbc5445-8l7jf" Dec 06 16:01:46 crc kubenswrapper[4813]: I1206 16:01:46.019400 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4400510-3d90-46c6-ba53-d964f310719a-config\") pod \"dnsmasq-dns-b8fbc5445-8l7jf\" (UID: \"b4400510-3d90-46c6-ba53-d964f310719a\") " pod="openstack/dnsmasq-dns-b8fbc5445-8l7jf" Dec 06 16:01:46 crc kubenswrapper[4813]: I1206 16:01:46.019511 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4400510-3d90-46c6-ba53-d964f310719a-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-8l7jf\" (UID: \"b4400510-3d90-46c6-ba53-d964f310719a\") " pod="openstack/dnsmasq-dns-b8fbc5445-8l7jf" Dec 06 16:01:46 crc kubenswrapper[4813]: I1206 16:01:46.019586 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4400510-3d90-46c6-ba53-d964f310719a-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-8l7jf\" (UID: \"b4400510-3d90-46c6-ba53-d964f310719a\") " pod="openstack/dnsmasq-dns-b8fbc5445-8l7jf" Dec 06 16:01:46 crc kubenswrapper[4813]: I1206 16:01:46.020165 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4400510-3d90-46c6-ba53-d964f310719a-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-8l7jf\" (UID: \"b4400510-3d90-46c6-ba53-d964f310719a\") " pod="openstack/dnsmasq-dns-b8fbc5445-8l7jf" Dec 06 16:01:46 crc kubenswrapper[4813]: I1206 16:01:46.051109 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2wzt\" (UniqueName: \"kubernetes.io/projected/b4400510-3d90-46c6-ba53-d964f310719a-kube-api-access-x2wzt\") pod \"dnsmasq-dns-b8fbc5445-8l7jf\" (UID: \"b4400510-3d90-46c6-ba53-d964f310719a\") " pod="openstack/dnsmasq-dns-b8fbc5445-8l7jf" Dec 06 16:01:46 crc kubenswrapper[4813]: I1206 16:01:46.122669 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 06 16:01:46 crc kubenswrapper[4813]: I1206 16:01:46.224679 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-8l7jf" Dec 06 16:01:46 crc kubenswrapper[4813]: I1206 16:01:46.331550 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 06 16:01:46 crc kubenswrapper[4813]: I1206 16:01:46.680071 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-8l7jf"] Dec 06 16:01:46 crc kubenswrapper[4813]: W1206 16:01:46.694681 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb4400510_3d90_46c6_ba53_d964f310719a.slice/crio-20ac4a2474e39665abda47620de8302af53ef196d3bc48519ff532816e47db72 WatchSource:0}: Error finding container 20ac4a2474e39665abda47620de8302af53ef196d3bc48519ff532816e47db72: Status 404 returned error can't find the container with id 20ac4a2474e39665abda47620de8302af53ef196d3bc48519ff532816e47db72 Dec 06 16:01:46 crc kubenswrapper[4813]: I1206 16:01:46.807947 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-kckf9" Dec 06 16:01:46 crc kubenswrapper[4813]: I1206 16:01:46.829494 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0497b076-9f37-46bb-a5dd-370f89e97b15-ovsdbserver-sb\") pod \"0497b076-9f37-46bb-a5dd-370f89e97b15\" (UID: \"0497b076-9f37-46bb-a5dd-370f89e97b15\") " Dec 06 16:01:46 crc kubenswrapper[4813]: I1206 16:01:46.829826 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lqhn\" (UniqueName: \"kubernetes.io/projected/0497b076-9f37-46bb-a5dd-370f89e97b15-kube-api-access-7lqhn\") pod \"0497b076-9f37-46bb-a5dd-370f89e97b15\" (UID: \"0497b076-9f37-46bb-a5dd-370f89e97b15\") " Dec 06 16:01:46 crc kubenswrapper[4813]: I1206 16:01:46.829851 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0497b076-9f37-46bb-a5dd-370f89e97b15-config\") pod \"0497b076-9f37-46bb-a5dd-370f89e97b15\" (UID: \"0497b076-9f37-46bb-a5dd-370f89e97b15\") " Dec 06 16:01:46 crc kubenswrapper[4813]: I1206 16:01:46.830031 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0497b076-9f37-46bb-a5dd-370f89e97b15-ovsdbserver-nb\") pod \"0497b076-9f37-46bb-a5dd-370f89e97b15\" (UID: \"0497b076-9f37-46bb-a5dd-370f89e97b15\") " Dec 06 16:01:46 crc kubenswrapper[4813]: I1206 16:01:46.830110 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0497b076-9f37-46bb-a5dd-370f89e97b15-dns-svc\") pod \"0497b076-9f37-46bb-a5dd-370f89e97b15\" (UID: \"0497b076-9f37-46bb-a5dd-370f89e97b15\") " Dec 06 16:01:46 crc kubenswrapper[4813]: I1206 16:01:46.861702 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0497b076-9f37-46bb-a5dd-370f89e97b15-kube-api-access-7lqhn" (OuterVolumeSpecName: "kube-api-access-7lqhn") pod "0497b076-9f37-46bb-a5dd-370f89e97b15" (UID: "0497b076-9f37-46bb-a5dd-370f89e97b15"). InnerVolumeSpecName "kube-api-access-7lqhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:01:46 crc kubenswrapper[4813]: I1206 16:01:46.914758 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0497b076-9f37-46bb-a5dd-370f89e97b15-config" (OuterVolumeSpecName: "config") pod "0497b076-9f37-46bb-a5dd-370f89e97b15" (UID: "0497b076-9f37-46bb-a5dd-370f89e97b15"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:01:46 crc kubenswrapper[4813]: I1206 16:01:46.919418 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0497b076-9f37-46bb-a5dd-370f89e97b15-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0497b076-9f37-46bb-a5dd-370f89e97b15" (UID: "0497b076-9f37-46bb-a5dd-370f89e97b15"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:01:46 crc kubenswrapper[4813]: I1206 16:01:46.919550 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0497b076-9f37-46bb-a5dd-370f89e97b15-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0497b076-9f37-46bb-a5dd-370f89e97b15" (UID: "0497b076-9f37-46bb-a5dd-370f89e97b15"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:01:46 crc kubenswrapper[4813]: I1206 16:01:46.926695 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0497b076-9f37-46bb-a5dd-370f89e97b15-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0497b076-9f37-46bb-a5dd-370f89e97b15" (UID: "0497b076-9f37-46bb-a5dd-370f89e97b15"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:01:46 crc kubenswrapper[4813]: I1206 16:01:46.935360 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0497b076-9f37-46bb-a5dd-370f89e97b15-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:46 crc kubenswrapper[4813]: I1206 16:01:46.935399 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0497b076-9f37-46bb-a5dd-370f89e97b15-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:46 crc kubenswrapper[4813]: I1206 16:01:46.935412 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lqhn\" (UniqueName: \"kubernetes.io/projected/0497b076-9f37-46bb-a5dd-370f89e97b15-kube-api-access-7lqhn\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:46 crc kubenswrapper[4813]: I1206 16:01:46.935421 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0497b076-9f37-46bb-a5dd-370f89e97b15-config\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:46 crc kubenswrapper[4813]: I1206 16:01:46.935430 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0497b076-9f37-46bb-a5dd-370f89e97b15-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.011944 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 06 16:01:47 crc kubenswrapper[4813]: E1206 16:01:47.012238 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0497b076-9f37-46bb-a5dd-370f89e97b15" containerName="dnsmasq-dns" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.012251 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0497b076-9f37-46bb-a5dd-370f89e97b15" containerName="dnsmasq-dns" Dec 06 16:01:47 crc kubenswrapper[4813]: E1206 16:01:47.012283 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0497b076-9f37-46bb-a5dd-370f89e97b15" containerName="init" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.012289 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0497b076-9f37-46bb-a5dd-370f89e97b15" containerName="init" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.012434 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0497b076-9f37-46bb-a5dd-370f89e97b15" containerName="dnsmasq-dns" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.016504 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-kckf9" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.019292 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-kckf9" event={"ID":"0497b076-9f37-46bb-a5dd-370f89e97b15","Type":"ContainerDied","Data":"00eb185060e50110c7144725493e065052bffe08d9c3075c9222f141ab00fcff"} Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.019334 4813 scope.go:117] "RemoveContainer" containerID="0b85051277ec559995330ca64a4036a37971d42cb977db03710e4f19d50d2401" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.019482 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.021987 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.022419 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.022484 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-4kjjv" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.022424 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.031443 4813 generic.go:334] "Generic (PLEG): container finished" podID="b4400510-3d90-46c6-ba53-d964f310719a" containerID="bd8496a393a15d54119a5e6ab856cd1a41772d40d077496f866f1d392899bf4b" exitCode=0 Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.031510 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-8l7jf" event={"ID":"b4400510-3d90-46c6-ba53-d964f310719a","Type":"ContainerDied","Data":"bd8496a393a15d54119a5e6ab856cd1a41772d40d077496f866f1d392899bf4b"} Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.031538 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-8l7jf" event={"ID":"b4400510-3d90-46c6-ba53-d964f310719a","Type":"ContainerStarted","Data":"20ac4a2474e39665abda47620de8302af53ef196d3bc48519ff532816e47db72"} Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.038050 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.042905 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"9614929b-2a14-4d49-8009-8f627fb000fd","Type":"ContainerStarted","Data":"d274dbd9db5f0156330b927c6d5844263ce8c57e5b31f58e8b5e82331abeb4b1"} Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.070565 4813 scope.go:117] "RemoveContainer" containerID="a9551fb18c6dd23ac19ebb75211a695bc01a9c410d898e26b1ea02ad90953574" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.087307 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-kckf9"] Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.097319 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-kckf9"] Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.138868 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tln5h\" (UniqueName: \"kubernetes.io/projected/4f571761-c8de-485f-89de-514fc0a66a55-kube-api-access-tln5h\") pod \"swift-storage-0\" (UID: \"4f571761-c8de-485f-89de-514fc0a66a55\") " pod="openstack/swift-storage-0" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.138947 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4f571761-c8de-485f-89de-514fc0a66a55-cache\") pod \"swift-storage-0\" (UID: \"4f571761-c8de-485f-89de-514fc0a66a55\") " pod="openstack/swift-storage-0" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.139008 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"4f571761-c8de-485f-89de-514fc0a66a55\") " pod="openstack/swift-storage-0" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.139098 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4f571761-c8de-485f-89de-514fc0a66a55-etc-swift\") pod \"swift-storage-0\" (UID: \"4f571761-c8de-485f-89de-514fc0a66a55\") " pod="openstack/swift-storage-0" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.139122 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4f571761-c8de-485f-89de-514fc0a66a55-lock\") pod \"swift-storage-0\" (UID: \"4f571761-c8de-485f-89de-514fc0a66a55\") " pod="openstack/swift-storage-0" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.240897 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4f571761-c8de-485f-89de-514fc0a66a55-lock\") pod \"swift-storage-0\" (UID: \"4f571761-c8de-485f-89de-514fc0a66a55\") " pod="openstack/swift-storage-0" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.241349 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tln5h\" (UniqueName: \"kubernetes.io/projected/4f571761-c8de-485f-89de-514fc0a66a55-kube-api-access-tln5h\") pod \"swift-storage-0\" (UID: \"4f571761-c8de-485f-89de-514fc0a66a55\") " pod="openstack/swift-storage-0" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.241372 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4f571761-c8de-485f-89de-514fc0a66a55-lock\") pod \"swift-storage-0\" (UID: \"4f571761-c8de-485f-89de-514fc0a66a55\") " pod="openstack/swift-storage-0" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.241386 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4f571761-c8de-485f-89de-514fc0a66a55-cache\") pod \"swift-storage-0\" (UID: \"4f571761-c8de-485f-89de-514fc0a66a55\") " pod="openstack/swift-storage-0" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.241444 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"4f571761-c8de-485f-89de-514fc0a66a55\") " pod="openstack/swift-storage-0" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.241508 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4f571761-c8de-485f-89de-514fc0a66a55-etc-swift\") pod \"swift-storage-0\" (UID: \"4f571761-c8de-485f-89de-514fc0a66a55\") " pod="openstack/swift-storage-0" Dec 06 16:01:47 crc kubenswrapper[4813]: E1206 16:01:47.241655 4813 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.241680 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4f571761-c8de-485f-89de-514fc0a66a55-cache\") pod \"swift-storage-0\" (UID: \"4f571761-c8de-485f-89de-514fc0a66a55\") " pod="openstack/swift-storage-0" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.241704 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"4f571761-c8de-485f-89de-514fc0a66a55\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/swift-storage-0" Dec 06 16:01:47 crc kubenswrapper[4813]: E1206 16:01:47.241683 4813 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 16:01:47 crc kubenswrapper[4813]: E1206 16:01:47.241875 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4f571761-c8de-485f-89de-514fc0a66a55-etc-swift podName:4f571761-c8de-485f-89de-514fc0a66a55 nodeName:}" failed. No retries permitted until 2025-12-06 16:01:47.741859084 +0000 UTC m=+947.632738660 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4f571761-c8de-485f-89de-514fc0a66a55-etc-swift") pod "swift-storage-0" (UID: "4f571761-c8de-485f-89de-514fc0a66a55") : configmap "swift-ring-files" not found Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.258107 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tln5h\" (UniqueName: \"kubernetes.io/projected/4f571761-c8de-485f-89de-514fc0a66a55-kube-api-access-tln5h\") pod \"swift-storage-0\" (UID: \"4f571761-c8de-485f-89de-514fc0a66a55\") " pod="openstack/swift-storage-0" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.261963 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"4f571761-c8de-485f-89de-514fc0a66a55\") " pod="openstack/swift-storage-0" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.554663 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-4dmqj"] Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.555835 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-4dmqj" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.557240 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.557542 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.557966 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.567544 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-4dmqj"] Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.622856 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-7pplm"] Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.623792 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-7pplm" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.634552 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-4dmqj"] Dec 06 16:01:47 crc kubenswrapper[4813]: E1206 16:01:47.638895 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-dclt9 ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-dclt9 ring-data-devices scripts swiftconf]: context canceled" pod="openstack/swift-ring-rebalance-4dmqj" podUID="20e380b2-81d0-470e-8475-dea6169fd6e4" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.646952 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/20e380b2-81d0-470e-8475-dea6169fd6e4-swiftconf\") pod \"swift-ring-rebalance-4dmqj\" (UID: \"20e380b2-81d0-470e-8475-dea6169fd6e4\") " pod="openstack/swift-ring-rebalance-4dmqj" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.652090 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-etc-swift\") pod \"swift-ring-rebalance-7pplm\" (UID: \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\") " pod="openstack/swift-ring-rebalance-7pplm" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.652225 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dclt9\" (UniqueName: \"kubernetes.io/projected/20e380b2-81d0-470e-8475-dea6169fd6e4-kube-api-access-dclt9\") pod \"swift-ring-rebalance-4dmqj\" (UID: \"20e380b2-81d0-470e-8475-dea6169fd6e4\") " pod="openstack/swift-ring-rebalance-4dmqj" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.652251 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-scripts\") pod \"swift-ring-rebalance-7pplm\" (UID: \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\") " pod="openstack/swift-ring-rebalance-7pplm" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.652460 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-combined-ca-bundle\") pod \"swift-ring-rebalance-7pplm\" (UID: \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\") " pod="openstack/swift-ring-rebalance-7pplm" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.653050 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/20e380b2-81d0-470e-8475-dea6169fd6e4-scripts\") pod \"swift-ring-rebalance-4dmqj\" (UID: \"20e380b2-81d0-470e-8475-dea6169fd6e4\") " pod="openstack/swift-ring-rebalance-4dmqj" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.653092 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/20e380b2-81d0-470e-8475-dea6169fd6e4-etc-swift\") pod \"swift-ring-rebalance-4dmqj\" (UID: \"20e380b2-81d0-470e-8475-dea6169fd6e4\") " pod="openstack/swift-ring-rebalance-4dmqj" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.653122 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/20e380b2-81d0-470e-8475-dea6169fd6e4-dispersionconf\") pod \"swift-ring-rebalance-4dmqj\" (UID: \"20e380b2-81d0-470e-8475-dea6169fd6e4\") " pod="openstack/swift-ring-rebalance-4dmqj" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.653152 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-swiftconf\") pod \"swift-ring-rebalance-7pplm\" (UID: \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\") " pod="openstack/swift-ring-rebalance-7pplm" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.653195 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-dispersionconf\") pod \"swift-ring-rebalance-7pplm\" (UID: \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\") " pod="openstack/swift-ring-rebalance-7pplm" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.653212 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9xg8\" (UniqueName: \"kubernetes.io/projected/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-kube-api-access-h9xg8\") pod \"swift-ring-rebalance-7pplm\" (UID: \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\") " pod="openstack/swift-ring-rebalance-7pplm" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.653235 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-ring-data-devices\") pod \"swift-ring-rebalance-7pplm\" (UID: \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\") " pod="openstack/swift-ring-rebalance-7pplm" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.653297 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20e380b2-81d0-470e-8475-dea6169fd6e4-combined-ca-bundle\") pod \"swift-ring-rebalance-4dmqj\" (UID: \"20e380b2-81d0-470e-8475-dea6169fd6e4\") " pod="openstack/swift-ring-rebalance-4dmqj" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.653314 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/20e380b2-81d0-470e-8475-dea6169fd6e4-ring-data-devices\") pod \"swift-ring-rebalance-4dmqj\" (UID: \"20e380b2-81d0-470e-8475-dea6169fd6e4\") " pod="openstack/swift-ring-rebalance-4dmqj" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.660460 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-7pplm"] Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.755019 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/20e380b2-81d0-470e-8475-dea6169fd6e4-scripts\") pod \"swift-ring-rebalance-4dmqj\" (UID: \"20e380b2-81d0-470e-8475-dea6169fd6e4\") " pod="openstack/swift-ring-rebalance-4dmqj" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.755061 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/20e380b2-81d0-470e-8475-dea6169fd6e4-etc-swift\") pod \"swift-ring-rebalance-4dmqj\" (UID: \"20e380b2-81d0-470e-8475-dea6169fd6e4\") " pod="openstack/swift-ring-rebalance-4dmqj" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.755088 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/20e380b2-81d0-470e-8475-dea6169fd6e4-dispersionconf\") pod \"swift-ring-rebalance-4dmqj\" (UID: \"20e380b2-81d0-470e-8475-dea6169fd6e4\") " pod="openstack/swift-ring-rebalance-4dmqj" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.755112 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-swiftconf\") pod \"swift-ring-rebalance-7pplm\" (UID: \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\") " pod="openstack/swift-ring-rebalance-7pplm" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.755135 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4f571761-c8de-485f-89de-514fc0a66a55-etc-swift\") pod \"swift-storage-0\" (UID: \"4f571761-c8de-485f-89de-514fc0a66a55\") " pod="openstack/swift-storage-0" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.755156 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-dispersionconf\") pod \"swift-ring-rebalance-7pplm\" (UID: \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\") " pod="openstack/swift-ring-rebalance-7pplm" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.755173 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9xg8\" (UniqueName: \"kubernetes.io/projected/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-kube-api-access-h9xg8\") pod \"swift-ring-rebalance-7pplm\" (UID: \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\") " pod="openstack/swift-ring-rebalance-7pplm" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.755207 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-ring-data-devices\") pod \"swift-ring-rebalance-7pplm\" (UID: \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\") " pod="openstack/swift-ring-rebalance-7pplm" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.755231 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20e380b2-81d0-470e-8475-dea6169fd6e4-combined-ca-bundle\") pod \"swift-ring-rebalance-4dmqj\" (UID: \"20e380b2-81d0-470e-8475-dea6169fd6e4\") " pod="openstack/swift-ring-rebalance-4dmqj" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.755269 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/20e380b2-81d0-470e-8475-dea6169fd6e4-ring-data-devices\") pod \"swift-ring-rebalance-4dmqj\" (UID: \"20e380b2-81d0-470e-8475-dea6169fd6e4\") " pod="openstack/swift-ring-rebalance-4dmqj" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.755303 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/20e380b2-81d0-470e-8475-dea6169fd6e4-swiftconf\") pod \"swift-ring-rebalance-4dmqj\" (UID: \"20e380b2-81d0-470e-8475-dea6169fd6e4\") " pod="openstack/swift-ring-rebalance-4dmqj" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.755320 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-etc-swift\") pod \"swift-ring-rebalance-7pplm\" (UID: \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\") " pod="openstack/swift-ring-rebalance-7pplm" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.755337 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dclt9\" (UniqueName: \"kubernetes.io/projected/20e380b2-81d0-470e-8475-dea6169fd6e4-kube-api-access-dclt9\") pod \"swift-ring-rebalance-4dmqj\" (UID: \"20e380b2-81d0-470e-8475-dea6169fd6e4\") " pod="openstack/swift-ring-rebalance-4dmqj" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.755352 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-scripts\") pod \"swift-ring-rebalance-7pplm\" (UID: \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\") " pod="openstack/swift-ring-rebalance-7pplm" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.755380 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-combined-ca-bundle\") pod \"swift-ring-rebalance-7pplm\" (UID: \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\") " pod="openstack/swift-ring-rebalance-7pplm" Dec 06 16:01:47 crc kubenswrapper[4813]: E1206 16:01:47.755822 4813 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 16:01:47 crc kubenswrapper[4813]: E1206 16:01:47.755846 4813 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 16:01:47 crc kubenswrapper[4813]: E1206 16:01:47.755907 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4f571761-c8de-485f-89de-514fc0a66a55-etc-swift podName:4f571761-c8de-485f-89de-514fc0a66a55 nodeName:}" failed. No retries permitted until 2025-12-06 16:01:48.75589082 +0000 UTC m=+948.646770396 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4f571761-c8de-485f-89de-514fc0a66a55-etc-swift") pod "swift-storage-0" (UID: "4f571761-c8de-485f-89de-514fc0a66a55") : configmap "swift-ring-files" not found Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.756896 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/20e380b2-81d0-470e-8475-dea6169fd6e4-etc-swift\") pod \"swift-ring-rebalance-4dmqj\" (UID: \"20e380b2-81d0-470e-8475-dea6169fd6e4\") " pod="openstack/swift-ring-rebalance-4dmqj" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.757389 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-ring-data-devices\") pod \"swift-ring-rebalance-7pplm\" (UID: \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\") " pod="openstack/swift-ring-rebalance-7pplm" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.757391 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-etc-swift\") pod \"swift-ring-rebalance-7pplm\" (UID: \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\") " pod="openstack/swift-ring-rebalance-7pplm" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.757822 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-scripts\") pod \"swift-ring-rebalance-7pplm\" (UID: \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\") " pod="openstack/swift-ring-rebalance-7pplm" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.758524 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/20e380b2-81d0-470e-8475-dea6169fd6e4-scripts\") pod \"swift-ring-rebalance-4dmqj\" (UID: \"20e380b2-81d0-470e-8475-dea6169fd6e4\") " pod="openstack/swift-ring-rebalance-4dmqj" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.761926 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-dispersionconf\") pod \"swift-ring-rebalance-7pplm\" (UID: \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\") " pod="openstack/swift-ring-rebalance-7pplm" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.762952 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/20e380b2-81d0-470e-8475-dea6169fd6e4-ring-data-devices\") pod \"swift-ring-rebalance-4dmqj\" (UID: \"20e380b2-81d0-470e-8475-dea6169fd6e4\") " pod="openstack/swift-ring-rebalance-4dmqj" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.763865 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/20e380b2-81d0-470e-8475-dea6169fd6e4-swiftconf\") pod \"swift-ring-rebalance-4dmqj\" (UID: \"20e380b2-81d0-470e-8475-dea6169fd6e4\") " pod="openstack/swift-ring-rebalance-4dmqj" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.764050 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-swiftconf\") pod \"swift-ring-rebalance-7pplm\" (UID: \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\") " pod="openstack/swift-ring-rebalance-7pplm" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.764149 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/20e380b2-81d0-470e-8475-dea6169fd6e4-dispersionconf\") pod \"swift-ring-rebalance-4dmqj\" (UID: \"20e380b2-81d0-470e-8475-dea6169fd6e4\") " pod="openstack/swift-ring-rebalance-4dmqj" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.764940 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-combined-ca-bundle\") pod \"swift-ring-rebalance-7pplm\" (UID: \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\") " pod="openstack/swift-ring-rebalance-7pplm" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.766712 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20e380b2-81d0-470e-8475-dea6169fd6e4-combined-ca-bundle\") pod \"swift-ring-rebalance-4dmqj\" (UID: \"20e380b2-81d0-470e-8475-dea6169fd6e4\") " pod="openstack/swift-ring-rebalance-4dmqj" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.775157 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dclt9\" (UniqueName: \"kubernetes.io/projected/20e380b2-81d0-470e-8475-dea6169fd6e4-kube-api-access-dclt9\") pod \"swift-ring-rebalance-4dmqj\" (UID: \"20e380b2-81d0-470e-8475-dea6169fd6e4\") " pod="openstack/swift-ring-rebalance-4dmqj" Dec 06 16:01:47 crc kubenswrapper[4813]: I1206 16:01:47.777365 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9xg8\" (UniqueName: \"kubernetes.io/projected/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-kube-api-access-h9xg8\") pod \"swift-ring-rebalance-7pplm\" (UID: \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\") " pod="openstack/swift-ring-rebalance-7pplm" Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.035386 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-7pplm" Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.058089 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-8l7jf" event={"ID":"b4400510-3d90-46c6-ba53-d964f310719a","Type":"ContainerStarted","Data":"11fcb034dbd10d0bf3b59810737248189753f8ccf7472663584df920cd838432"} Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.058230 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-8l7jf" Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.060207 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-4dmqj" Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.060233 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"9614929b-2a14-4d49-8009-8f627fb000fd","Type":"ContainerStarted","Data":"13a88611ef5e3ec25f9fbfd26fbb8e1682ef1a1948e2ea0f6865afdf906fe6ff"} Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.060273 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"9614929b-2a14-4d49-8009-8f627fb000fd","Type":"ContainerStarted","Data":"8da20c0ea62ab044d72a7a105293bb35f6794e37818dc1d0b4cc8f1edc59bb60"} Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.060443 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.069892 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-4dmqj" Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.103607 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.870379231 podStartE2EDuration="3.103587333s" podCreationTimestamp="2025-12-06 16:01:45 +0000 UTC" firstStartedPulling="2025-12-06 16:01:46.358839391 +0000 UTC m=+946.249718967" lastFinishedPulling="2025-12-06 16:01:47.592047483 +0000 UTC m=+947.482927069" observedRunningTime="2025-12-06 16:01:48.102290679 +0000 UTC m=+947.993170255" watchObservedRunningTime="2025-12-06 16:01:48.103587333 +0000 UTC m=+947.994466919" Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.109471 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-8l7jf" podStartSLOduration=3.109457379 podStartE2EDuration="3.109457379s" podCreationTimestamp="2025-12-06 16:01:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:01:48.07625824 +0000 UTC m=+947.967137856" watchObservedRunningTime="2025-12-06 16:01:48.109457379 +0000 UTC m=+948.000336975" Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.159209 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/20e380b2-81d0-470e-8475-dea6169fd6e4-ring-data-devices\") pod \"20e380b2-81d0-470e-8475-dea6169fd6e4\" (UID: \"20e380b2-81d0-470e-8475-dea6169fd6e4\") " Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.159488 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/20e380b2-81d0-470e-8475-dea6169fd6e4-dispersionconf\") pod \"20e380b2-81d0-470e-8475-dea6169fd6e4\" (UID: \"20e380b2-81d0-470e-8475-dea6169fd6e4\") " Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.159534 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/20e380b2-81d0-470e-8475-dea6169fd6e4-swiftconf\") pod \"20e380b2-81d0-470e-8475-dea6169fd6e4\" (UID: \"20e380b2-81d0-470e-8475-dea6169fd6e4\") " Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.159554 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20e380b2-81d0-470e-8475-dea6169fd6e4-combined-ca-bundle\") pod \"20e380b2-81d0-470e-8475-dea6169fd6e4\" (UID: \"20e380b2-81d0-470e-8475-dea6169fd6e4\") " Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.159579 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dclt9\" (UniqueName: \"kubernetes.io/projected/20e380b2-81d0-470e-8475-dea6169fd6e4-kube-api-access-dclt9\") pod \"20e380b2-81d0-470e-8475-dea6169fd6e4\" (UID: \"20e380b2-81d0-470e-8475-dea6169fd6e4\") " Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.159596 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/20e380b2-81d0-470e-8475-dea6169fd6e4-scripts\") pod \"20e380b2-81d0-470e-8475-dea6169fd6e4\" (UID: \"20e380b2-81d0-470e-8475-dea6169fd6e4\") " Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.159623 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/20e380b2-81d0-470e-8475-dea6169fd6e4-etc-swift\") pod \"20e380b2-81d0-470e-8475-dea6169fd6e4\" (UID: \"20e380b2-81d0-470e-8475-dea6169fd6e4\") " Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.160997 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20e380b2-81d0-470e-8475-dea6169fd6e4-scripts" (OuterVolumeSpecName: "scripts") pod "20e380b2-81d0-470e-8475-dea6169fd6e4" (UID: "20e380b2-81d0-470e-8475-dea6169fd6e4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.161193 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20e380b2-81d0-470e-8475-dea6169fd6e4-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "20e380b2-81d0-470e-8475-dea6169fd6e4" (UID: "20e380b2-81d0-470e-8475-dea6169fd6e4"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.161901 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20e380b2-81d0-470e-8475-dea6169fd6e4-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "20e380b2-81d0-470e-8475-dea6169fd6e4" (UID: "20e380b2-81d0-470e-8475-dea6169fd6e4"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.165029 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20e380b2-81d0-470e-8475-dea6169fd6e4-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "20e380b2-81d0-470e-8475-dea6169fd6e4" (UID: "20e380b2-81d0-470e-8475-dea6169fd6e4"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.165090 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20e380b2-81d0-470e-8475-dea6169fd6e4-kube-api-access-dclt9" (OuterVolumeSpecName: "kube-api-access-dclt9") pod "20e380b2-81d0-470e-8475-dea6169fd6e4" (UID: "20e380b2-81d0-470e-8475-dea6169fd6e4"). InnerVolumeSpecName "kube-api-access-dclt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.165422 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20e380b2-81d0-470e-8475-dea6169fd6e4-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "20e380b2-81d0-470e-8475-dea6169fd6e4" (UID: "20e380b2-81d0-470e-8475-dea6169fd6e4"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.166481 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20e380b2-81d0-470e-8475-dea6169fd6e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "20e380b2-81d0-470e-8475-dea6169fd6e4" (UID: "20e380b2-81d0-470e-8475-dea6169fd6e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.191785 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.261997 4813 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/20e380b2-81d0-470e-8475-dea6169fd6e4-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.262033 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20e380b2-81d0-470e-8475-dea6169fd6e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.262048 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dclt9\" (UniqueName: \"kubernetes.io/projected/20e380b2-81d0-470e-8475-dea6169fd6e4-kube-api-access-dclt9\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.262061 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/20e380b2-81d0-470e-8475-dea6169fd6e4-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.262072 4813 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/20e380b2-81d0-470e-8475-dea6169fd6e4-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.262084 4813 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/20e380b2-81d0-470e-8475-dea6169fd6e4-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.262096 4813 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/20e380b2-81d0-470e-8475-dea6169fd6e4-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.282649 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.455675 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-7pplm"] Dec 06 16:01:48 crc kubenswrapper[4813]: W1206 16:01:48.458741 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf67c4c64_2c67_46bc_af93_6ca6ac4f4814.slice/crio-feae26864325a76ec28c1c9f07e5b42eb3ad928af4fb1104f03b55bdcbb5682e WatchSource:0}: Error finding container feae26864325a76ec28c1c9f07e5b42eb3ad928af4fb1104f03b55bdcbb5682e: Status 404 returned error can't find the container with id feae26864325a76ec28c1c9f07e5b42eb3ad928af4fb1104f03b55bdcbb5682e Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.498312 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0497b076-9f37-46bb-a5dd-370f89e97b15" path="/var/lib/kubelet/pods/0497b076-9f37-46bb-a5dd-370f89e97b15/volumes" Dec 06 16:01:48 crc kubenswrapper[4813]: I1206 16:01:48.774353 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4f571761-c8de-485f-89de-514fc0a66a55-etc-swift\") pod \"swift-storage-0\" (UID: \"4f571761-c8de-485f-89de-514fc0a66a55\") " pod="openstack/swift-storage-0" Dec 06 16:01:48 crc kubenswrapper[4813]: E1206 16:01:48.774630 4813 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 16:01:48 crc kubenswrapper[4813]: E1206 16:01:48.774855 4813 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 16:01:48 crc kubenswrapper[4813]: E1206 16:01:48.774946 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4f571761-c8de-485f-89de-514fc0a66a55-etc-swift podName:4f571761-c8de-485f-89de-514fc0a66a55 nodeName:}" failed. No retries permitted until 2025-12-06 16:01:50.774917222 +0000 UTC m=+950.665796838 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4f571761-c8de-485f-89de-514fc0a66a55-etc-swift") pod "swift-storage-0" (UID: "4f571761-c8de-485f-89de-514fc0a66a55") : configmap "swift-ring-files" not found Dec 06 16:01:49 crc kubenswrapper[4813]: I1206 16:01:49.069245 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-7pplm" event={"ID":"f67c4c64-2c67-46bc-af93-6ca6ac4f4814","Type":"ContainerStarted","Data":"feae26864325a76ec28c1c9f07e5b42eb3ad928af4fb1104f03b55bdcbb5682e"} Dec 06 16:01:49 crc kubenswrapper[4813]: I1206 16:01:49.069371 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-4dmqj" Dec 06 16:01:49 crc kubenswrapper[4813]: I1206 16:01:49.128292 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-4dmqj"] Dec 06 16:01:49 crc kubenswrapper[4813]: I1206 16:01:49.134351 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-4dmqj"] Dec 06 16:01:49 crc kubenswrapper[4813]: I1206 16:01:49.214456 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-a434-account-create-update-5rxjv"] Dec 06 16:01:49 crc kubenswrapper[4813]: I1206 16:01:49.215638 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a434-account-create-update-5rxjv" Dec 06 16:01:49 crc kubenswrapper[4813]: I1206 16:01:49.217457 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-a434-account-create-update-5rxjv"] Dec 06 16:01:49 crc kubenswrapper[4813]: I1206 16:01:49.218918 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 06 16:01:49 crc kubenswrapper[4813]: I1206 16:01:49.260011 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-lwz48"] Dec 06 16:01:49 crc kubenswrapper[4813]: I1206 16:01:49.261081 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-lwz48" Dec 06 16:01:49 crc kubenswrapper[4813]: I1206 16:01:49.266671 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-lwz48"] Dec 06 16:01:49 crc kubenswrapper[4813]: I1206 16:01:49.385323 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/854df38a-5bef-4b2c-acdc-4ca7577bc1de-operator-scripts\") pod \"glance-a434-account-create-update-5rxjv\" (UID: \"854df38a-5bef-4b2c-acdc-4ca7577bc1de\") " pod="openstack/glance-a434-account-create-update-5rxjv" Dec 06 16:01:49 crc kubenswrapper[4813]: I1206 16:01:49.385412 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5q2nc\" (UniqueName: \"kubernetes.io/projected/66c65f92-2424-4d91-9824-9a6c00bf3dbf-kube-api-access-5q2nc\") pod \"glance-db-create-lwz48\" (UID: \"66c65f92-2424-4d91-9824-9a6c00bf3dbf\") " pod="openstack/glance-db-create-lwz48" Dec 06 16:01:49 crc kubenswrapper[4813]: I1206 16:01:49.385465 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/66c65f92-2424-4d91-9824-9a6c00bf3dbf-operator-scripts\") pod \"glance-db-create-lwz48\" (UID: \"66c65f92-2424-4d91-9824-9a6c00bf3dbf\") " pod="openstack/glance-db-create-lwz48" Dec 06 16:01:49 crc kubenswrapper[4813]: I1206 16:01:49.385983 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5h5lk\" (UniqueName: \"kubernetes.io/projected/854df38a-5bef-4b2c-acdc-4ca7577bc1de-kube-api-access-5h5lk\") pod \"glance-a434-account-create-update-5rxjv\" (UID: \"854df38a-5bef-4b2c-acdc-4ca7577bc1de\") " pod="openstack/glance-a434-account-create-update-5rxjv" Dec 06 16:01:49 crc kubenswrapper[4813]: I1206 16:01:49.487506 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5h5lk\" (UniqueName: \"kubernetes.io/projected/854df38a-5bef-4b2c-acdc-4ca7577bc1de-kube-api-access-5h5lk\") pod \"glance-a434-account-create-update-5rxjv\" (UID: \"854df38a-5bef-4b2c-acdc-4ca7577bc1de\") " pod="openstack/glance-a434-account-create-update-5rxjv" Dec 06 16:01:49 crc kubenswrapper[4813]: I1206 16:01:49.487555 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/854df38a-5bef-4b2c-acdc-4ca7577bc1de-operator-scripts\") pod \"glance-a434-account-create-update-5rxjv\" (UID: \"854df38a-5bef-4b2c-acdc-4ca7577bc1de\") " pod="openstack/glance-a434-account-create-update-5rxjv" Dec 06 16:01:49 crc kubenswrapper[4813]: I1206 16:01:49.487633 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5q2nc\" (UniqueName: \"kubernetes.io/projected/66c65f92-2424-4d91-9824-9a6c00bf3dbf-kube-api-access-5q2nc\") pod \"glance-db-create-lwz48\" (UID: \"66c65f92-2424-4d91-9824-9a6c00bf3dbf\") " pod="openstack/glance-db-create-lwz48" Dec 06 16:01:49 crc kubenswrapper[4813]: I1206 16:01:49.487734 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/66c65f92-2424-4d91-9824-9a6c00bf3dbf-operator-scripts\") pod \"glance-db-create-lwz48\" (UID: \"66c65f92-2424-4d91-9824-9a6c00bf3dbf\") " pod="openstack/glance-db-create-lwz48" Dec 06 16:01:49 crc kubenswrapper[4813]: I1206 16:01:49.488280 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/854df38a-5bef-4b2c-acdc-4ca7577bc1de-operator-scripts\") pod \"glance-a434-account-create-update-5rxjv\" (UID: \"854df38a-5bef-4b2c-acdc-4ca7577bc1de\") " pod="openstack/glance-a434-account-create-update-5rxjv" Dec 06 16:01:49 crc kubenswrapper[4813]: I1206 16:01:49.491400 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/66c65f92-2424-4d91-9824-9a6c00bf3dbf-operator-scripts\") pod \"glance-db-create-lwz48\" (UID: \"66c65f92-2424-4d91-9824-9a6c00bf3dbf\") " pod="openstack/glance-db-create-lwz48" Dec 06 16:01:49 crc kubenswrapper[4813]: I1206 16:01:49.505834 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5q2nc\" (UniqueName: \"kubernetes.io/projected/66c65f92-2424-4d91-9824-9a6c00bf3dbf-kube-api-access-5q2nc\") pod \"glance-db-create-lwz48\" (UID: \"66c65f92-2424-4d91-9824-9a6c00bf3dbf\") " pod="openstack/glance-db-create-lwz48" Dec 06 16:01:49 crc kubenswrapper[4813]: I1206 16:01:49.515786 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5h5lk\" (UniqueName: \"kubernetes.io/projected/854df38a-5bef-4b2c-acdc-4ca7577bc1de-kube-api-access-5h5lk\") pod \"glance-a434-account-create-update-5rxjv\" (UID: \"854df38a-5bef-4b2c-acdc-4ca7577bc1de\") " pod="openstack/glance-a434-account-create-update-5rxjv" Dec 06 16:01:49 crc kubenswrapper[4813]: I1206 16:01:49.541529 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a434-account-create-update-5rxjv" Dec 06 16:01:49 crc kubenswrapper[4813]: I1206 16:01:49.579659 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-lwz48" Dec 06 16:01:50 crc kubenswrapper[4813]: I1206 16:01:50.010513 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-a434-account-create-update-5rxjv"] Dec 06 16:01:50 crc kubenswrapper[4813]: I1206 16:01:50.097000 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-lwz48"] Dec 06 16:01:50 crc kubenswrapper[4813]: I1206 16:01:50.512542 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20e380b2-81d0-470e-8475-dea6169fd6e4" path="/var/lib/kubelet/pods/20e380b2-81d0-470e-8475-dea6169fd6e4/volumes" Dec 06 16:01:50 crc kubenswrapper[4813]: I1206 16:01:50.811571 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4f571761-c8de-485f-89de-514fc0a66a55-etc-swift\") pod \"swift-storage-0\" (UID: \"4f571761-c8de-485f-89de-514fc0a66a55\") " pod="openstack/swift-storage-0" Dec 06 16:01:50 crc kubenswrapper[4813]: E1206 16:01:50.811799 4813 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 16:01:50 crc kubenswrapper[4813]: E1206 16:01:50.811848 4813 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 16:01:50 crc kubenswrapper[4813]: E1206 16:01:50.811967 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4f571761-c8de-485f-89de-514fc0a66a55-etc-swift podName:4f571761-c8de-485f-89de-514fc0a66a55 nodeName:}" failed. No retries permitted until 2025-12-06 16:01:54.81194982 +0000 UTC m=+954.702829396 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4f571761-c8de-485f-89de-514fc0a66a55-etc-swift") pod "swift-storage-0" (UID: "4f571761-c8de-485f-89de-514fc0a66a55") : configmap "swift-ring-files" not found Dec 06 16:01:51 crc kubenswrapper[4813]: I1206 16:01:51.086744 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a434-account-create-update-5rxjv" event={"ID":"854df38a-5bef-4b2c-acdc-4ca7577bc1de","Type":"ContainerStarted","Data":"d0487c8f9062dd074487c1690aad9e1450b105b09a47f33a3ea823e92727d6ce"} Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.274749 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-c96d-account-create-update-ldmnt"] Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.277141 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c96d-account-create-update-ldmnt" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.279030 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.291892 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-c96d-account-create-update-ldmnt"] Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.342990 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-qcgfq"] Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.344007 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-qcgfq" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.391112 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-2d5d-account-create-update-2pzsb"] Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.392298 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-2d5d-account-create-update-2pzsb" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.394093 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.397821 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-qcgfq"] Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.409649 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tnzt\" (UniqueName: \"kubernetes.io/projected/d882403f-89d4-4bef-b2a0-7f1134e4b735-kube-api-access-9tnzt\") pod \"keystone-c96d-account-create-update-ldmnt\" (UID: \"d882403f-89d4-4bef-b2a0-7f1134e4b735\") " pod="openstack/keystone-c96d-account-create-update-ldmnt" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.409692 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfz7l\" (UniqueName: \"kubernetes.io/projected/6805af73-82e1-4e15-b867-d02abc99d5ee-kube-api-access-qfz7l\") pod \"keystone-db-create-qcgfq\" (UID: \"6805af73-82e1-4e15-b867-d02abc99d5ee\") " pod="openstack/keystone-db-create-qcgfq" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.409736 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6805af73-82e1-4e15-b867-d02abc99d5ee-operator-scripts\") pod \"keystone-db-create-qcgfq\" (UID: \"6805af73-82e1-4e15-b867-d02abc99d5ee\") " pod="openstack/keystone-db-create-qcgfq" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.409794 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d882403f-89d4-4bef-b2a0-7f1134e4b735-operator-scripts\") pod \"keystone-c96d-account-create-update-ldmnt\" (UID: \"d882403f-89d4-4bef-b2a0-7f1134e4b735\") " pod="openstack/keystone-c96d-account-create-update-ldmnt" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.412674 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-2d5d-account-create-update-2pzsb"] Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.440593 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-wt55k"] Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.443386 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wt55k" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.462200 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-wt55k"] Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.516139 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tnzt\" (UniqueName: \"kubernetes.io/projected/d882403f-89d4-4bef-b2a0-7f1134e4b735-kube-api-access-9tnzt\") pod \"keystone-c96d-account-create-update-ldmnt\" (UID: \"d882403f-89d4-4bef-b2a0-7f1134e4b735\") " pod="openstack/keystone-c96d-account-create-update-ldmnt" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.516193 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfz7l\" (UniqueName: \"kubernetes.io/projected/6805af73-82e1-4e15-b867-d02abc99d5ee-kube-api-access-qfz7l\") pod \"keystone-db-create-qcgfq\" (UID: \"6805af73-82e1-4e15-b867-d02abc99d5ee\") " pod="openstack/keystone-db-create-qcgfq" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.516224 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6805af73-82e1-4e15-b867-d02abc99d5ee-operator-scripts\") pod \"keystone-db-create-qcgfq\" (UID: \"6805af73-82e1-4e15-b867-d02abc99d5ee\") " pod="openstack/keystone-db-create-qcgfq" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.516272 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7dd308a-25f5-4746-b850-d4a2430c6b20-operator-scripts\") pod \"placement-2d5d-account-create-update-2pzsb\" (UID: \"e7dd308a-25f5-4746-b850-d4a2430c6b20\") " pod="openstack/placement-2d5d-account-create-update-2pzsb" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.516321 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wb7nl\" (UniqueName: \"kubernetes.io/projected/e7dd308a-25f5-4746-b850-d4a2430c6b20-kube-api-access-wb7nl\") pod \"placement-2d5d-account-create-update-2pzsb\" (UID: \"e7dd308a-25f5-4746-b850-d4a2430c6b20\") " pod="openstack/placement-2d5d-account-create-update-2pzsb" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.516352 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d882403f-89d4-4bef-b2a0-7f1134e4b735-operator-scripts\") pod \"keystone-c96d-account-create-update-ldmnt\" (UID: \"d882403f-89d4-4bef-b2a0-7f1134e4b735\") " pod="openstack/keystone-c96d-account-create-update-ldmnt" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.516401 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee373829-88aa-427e-aedb-af7410b7b7a3-operator-scripts\") pod \"placement-db-create-wt55k\" (UID: \"ee373829-88aa-427e-aedb-af7410b7b7a3\") " pod="openstack/placement-db-create-wt55k" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.516450 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwzqm\" (UniqueName: \"kubernetes.io/projected/ee373829-88aa-427e-aedb-af7410b7b7a3-kube-api-access-zwzqm\") pod \"placement-db-create-wt55k\" (UID: \"ee373829-88aa-427e-aedb-af7410b7b7a3\") " pod="openstack/placement-db-create-wt55k" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.517057 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6805af73-82e1-4e15-b867-d02abc99d5ee-operator-scripts\") pod \"keystone-db-create-qcgfq\" (UID: \"6805af73-82e1-4e15-b867-d02abc99d5ee\") " pod="openstack/keystone-db-create-qcgfq" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.517170 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d882403f-89d4-4bef-b2a0-7f1134e4b735-operator-scripts\") pod \"keystone-c96d-account-create-update-ldmnt\" (UID: \"d882403f-89d4-4bef-b2a0-7f1134e4b735\") " pod="openstack/keystone-c96d-account-create-update-ldmnt" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.538402 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfz7l\" (UniqueName: \"kubernetes.io/projected/6805af73-82e1-4e15-b867-d02abc99d5ee-kube-api-access-qfz7l\") pod \"keystone-db-create-qcgfq\" (UID: \"6805af73-82e1-4e15-b867-d02abc99d5ee\") " pod="openstack/keystone-db-create-qcgfq" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.547700 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tnzt\" (UniqueName: \"kubernetes.io/projected/d882403f-89d4-4bef-b2a0-7f1134e4b735-kube-api-access-9tnzt\") pod \"keystone-c96d-account-create-update-ldmnt\" (UID: \"d882403f-89d4-4bef-b2a0-7f1134e4b735\") " pod="openstack/keystone-c96d-account-create-update-ldmnt" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.595676 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c96d-account-create-update-ldmnt" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.617654 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7dd308a-25f5-4746-b850-d4a2430c6b20-operator-scripts\") pod \"placement-2d5d-account-create-update-2pzsb\" (UID: \"e7dd308a-25f5-4746-b850-d4a2430c6b20\") " pod="openstack/placement-2d5d-account-create-update-2pzsb" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.617803 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wb7nl\" (UniqueName: \"kubernetes.io/projected/e7dd308a-25f5-4746-b850-d4a2430c6b20-kube-api-access-wb7nl\") pod \"placement-2d5d-account-create-update-2pzsb\" (UID: \"e7dd308a-25f5-4746-b850-d4a2430c6b20\") " pod="openstack/placement-2d5d-account-create-update-2pzsb" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.617907 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee373829-88aa-427e-aedb-af7410b7b7a3-operator-scripts\") pod \"placement-db-create-wt55k\" (UID: \"ee373829-88aa-427e-aedb-af7410b7b7a3\") " pod="openstack/placement-db-create-wt55k" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.618016 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwzqm\" (UniqueName: \"kubernetes.io/projected/ee373829-88aa-427e-aedb-af7410b7b7a3-kube-api-access-zwzqm\") pod \"placement-db-create-wt55k\" (UID: \"ee373829-88aa-427e-aedb-af7410b7b7a3\") " pod="openstack/placement-db-create-wt55k" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.619150 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7dd308a-25f5-4746-b850-d4a2430c6b20-operator-scripts\") pod \"placement-2d5d-account-create-update-2pzsb\" (UID: \"e7dd308a-25f5-4746-b850-d4a2430c6b20\") " pod="openstack/placement-2d5d-account-create-update-2pzsb" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.619339 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee373829-88aa-427e-aedb-af7410b7b7a3-operator-scripts\") pod \"placement-db-create-wt55k\" (UID: \"ee373829-88aa-427e-aedb-af7410b7b7a3\") " pod="openstack/placement-db-create-wt55k" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.635106 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwzqm\" (UniqueName: \"kubernetes.io/projected/ee373829-88aa-427e-aedb-af7410b7b7a3-kube-api-access-zwzqm\") pod \"placement-db-create-wt55k\" (UID: \"ee373829-88aa-427e-aedb-af7410b7b7a3\") " pod="openstack/placement-db-create-wt55k" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.635881 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wb7nl\" (UniqueName: \"kubernetes.io/projected/e7dd308a-25f5-4746-b850-d4a2430c6b20-kube-api-access-wb7nl\") pod \"placement-2d5d-account-create-update-2pzsb\" (UID: \"e7dd308a-25f5-4746-b850-d4a2430c6b20\") " pod="openstack/placement-2d5d-account-create-update-2pzsb" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.685557 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-qcgfq" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.741367 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-2d5d-account-create-update-2pzsb" Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.822867 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4f571761-c8de-485f-89de-514fc0a66a55-etc-swift\") pod \"swift-storage-0\" (UID: \"4f571761-c8de-485f-89de-514fc0a66a55\") " pod="openstack/swift-storage-0" Dec 06 16:01:54 crc kubenswrapper[4813]: E1206 16:01:54.823077 4813 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 16:01:54 crc kubenswrapper[4813]: E1206 16:01:54.823104 4813 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 16:01:54 crc kubenswrapper[4813]: E1206 16:01:54.823153 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4f571761-c8de-485f-89de-514fc0a66a55-etc-swift podName:4f571761-c8de-485f-89de-514fc0a66a55 nodeName:}" failed. No retries permitted until 2025-12-06 16:02:02.823133681 +0000 UTC m=+962.714013257 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4f571761-c8de-485f-89de-514fc0a66a55-etc-swift") pod "swift-storage-0" (UID: "4f571761-c8de-485f-89de-514fc0a66a55") : configmap "swift-ring-files" not found Dec 06 16:01:54 crc kubenswrapper[4813]: I1206 16:01:54.916268 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wt55k" Dec 06 16:01:55 crc kubenswrapper[4813]: I1206 16:01:55.077032 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-c96d-account-create-update-ldmnt"] Dec 06 16:01:55 crc kubenswrapper[4813]: W1206 16:01:55.083143 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd882403f_89d4_4bef_b2a0_7f1134e4b735.slice/crio-e0b38d8afd058fec9102e0f80619790089b4597cc928a5c12e048fb6b9723417 WatchSource:0}: Error finding container e0b38d8afd058fec9102e0f80619790089b4597cc928a5c12e048fb6b9723417: Status 404 returned error can't find the container with id e0b38d8afd058fec9102e0f80619790089b4597cc928a5c12e048fb6b9723417 Dec 06 16:01:55 crc kubenswrapper[4813]: I1206 16:01:55.133366 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-7pplm" event={"ID":"f67c4c64-2c67-46bc-af93-6ca6ac4f4814","Type":"ContainerStarted","Data":"bd9877744dc4c39dbe47b728f189edbcd112064748429eec348a62b3b78bc0e9"} Dec 06 16:01:55 crc kubenswrapper[4813]: I1206 16:01:55.135543 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c96d-account-create-update-ldmnt" event={"ID":"d882403f-89d4-4bef-b2a0-7f1134e4b735","Type":"ContainerStarted","Data":"e0b38d8afd058fec9102e0f80619790089b4597cc928a5c12e048fb6b9723417"} Dec 06 16:01:55 crc kubenswrapper[4813]: I1206 16:01:55.142698 4813 generic.go:334] "Generic (PLEG): container finished" podID="854df38a-5bef-4b2c-acdc-4ca7577bc1de" containerID="bf1a2528259ee03d13821c200c160bd9dc563fcfb453efc86aee5a6e923d1a7c" exitCode=0 Dec 06 16:01:55 crc kubenswrapper[4813]: I1206 16:01:55.142766 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a434-account-create-update-5rxjv" event={"ID":"854df38a-5bef-4b2c-acdc-4ca7577bc1de","Type":"ContainerDied","Data":"bf1a2528259ee03d13821c200c160bd9dc563fcfb453efc86aee5a6e923d1a7c"} Dec 06 16:01:55 crc kubenswrapper[4813]: I1206 16:01:55.152424 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-7pplm" podStartSLOduration=2.166599 podStartE2EDuration="8.152405447s" podCreationTimestamp="2025-12-06 16:01:47 +0000 UTC" firstStartedPulling="2025-12-06 16:01:48.461053275 +0000 UTC m=+948.351932851" lastFinishedPulling="2025-12-06 16:01:54.446859732 +0000 UTC m=+954.337739298" observedRunningTime="2025-12-06 16:01:55.145765651 +0000 UTC m=+955.036645227" watchObservedRunningTime="2025-12-06 16:01:55.152405447 +0000 UTC m=+955.043285023" Dec 06 16:01:55 crc kubenswrapper[4813]: I1206 16:01:55.154196 4813 generic.go:334] "Generic (PLEG): container finished" podID="66c65f92-2424-4d91-9824-9a6c00bf3dbf" containerID="e9b6e23ad4f2a2290ab5e5d6deea14532e1fe49412cd92438393dc10cad29994" exitCode=0 Dec 06 16:01:55 crc kubenswrapper[4813]: I1206 16:01:55.154236 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-lwz48" event={"ID":"66c65f92-2424-4d91-9824-9a6c00bf3dbf","Type":"ContainerDied","Data":"e9b6e23ad4f2a2290ab5e5d6deea14532e1fe49412cd92438393dc10cad29994"} Dec 06 16:01:55 crc kubenswrapper[4813]: I1206 16:01:55.154276 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-lwz48" event={"ID":"66c65f92-2424-4d91-9824-9a6c00bf3dbf","Type":"ContainerStarted","Data":"e8c6183bc681ff680bdbd9e98d28ebcd4587072c70740e476d7cd61fe6df1583"} Dec 06 16:01:55 crc kubenswrapper[4813]: I1206 16:01:55.237141 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-2d5d-account-create-update-2pzsb"] Dec 06 16:01:55 crc kubenswrapper[4813]: W1206 16:01:55.243463 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6805af73_82e1_4e15_b867_d02abc99d5ee.slice/crio-1434cc76fe9a6998b70c27aa9043e85711adeccfeef1cb0a4f740e5d17567576 WatchSource:0}: Error finding container 1434cc76fe9a6998b70c27aa9043e85711adeccfeef1cb0a4f740e5d17567576: Status 404 returned error can't find the container with id 1434cc76fe9a6998b70c27aa9043e85711adeccfeef1cb0a4f740e5d17567576 Dec 06 16:01:55 crc kubenswrapper[4813]: I1206 16:01:55.243742 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-qcgfq"] Dec 06 16:01:55 crc kubenswrapper[4813]: W1206 16:01:55.246275 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7dd308a_25f5_4746_b850_d4a2430c6b20.slice/crio-002a461e040d36b6f9eed06e31d1b247e72e3e68891b2bb67fa52cacd2898cef WatchSource:0}: Error finding container 002a461e040d36b6f9eed06e31d1b247e72e3e68891b2bb67fa52cacd2898cef: Status 404 returned error can't find the container with id 002a461e040d36b6f9eed06e31d1b247e72e3e68891b2bb67fa52cacd2898cef Dec 06 16:01:55 crc kubenswrapper[4813]: I1206 16:01:55.371553 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-wt55k"] Dec 06 16:01:55 crc kubenswrapper[4813]: W1206 16:01:55.395681 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee373829_88aa_427e_aedb_af7410b7b7a3.slice/crio-727d7c8fd4edce45ba7d9bc44fc91d9ee65c68cfebe3080b44f9db1d569a0c75 WatchSource:0}: Error finding container 727d7c8fd4edce45ba7d9bc44fc91d9ee65c68cfebe3080b44f9db1d569a0c75: Status 404 returned error can't find the container with id 727d7c8fd4edce45ba7d9bc44fc91d9ee65c68cfebe3080b44f9db1d569a0c75 Dec 06 16:01:56 crc kubenswrapper[4813]: I1206 16:01:56.168658 4813 generic.go:334] "Generic (PLEG): container finished" podID="6805af73-82e1-4e15-b867-d02abc99d5ee" containerID="af1becfbd82c4ea9d74467e0babd6618407bf7ef7dc9668d2318f06ee35ae8c2" exitCode=0 Dec 06 16:01:56 crc kubenswrapper[4813]: I1206 16:01:56.168775 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-qcgfq" event={"ID":"6805af73-82e1-4e15-b867-d02abc99d5ee","Type":"ContainerDied","Data":"af1becfbd82c4ea9d74467e0babd6618407bf7ef7dc9668d2318f06ee35ae8c2"} Dec 06 16:01:56 crc kubenswrapper[4813]: I1206 16:01:56.168836 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-qcgfq" event={"ID":"6805af73-82e1-4e15-b867-d02abc99d5ee","Type":"ContainerStarted","Data":"1434cc76fe9a6998b70c27aa9043e85711adeccfeef1cb0a4f740e5d17567576"} Dec 06 16:01:56 crc kubenswrapper[4813]: I1206 16:01:56.174074 4813 generic.go:334] "Generic (PLEG): container finished" podID="d882403f-89d4-4bef-b2a0-7f1134e4b735" containerID="f5b499d0437ecc1d705d99c428fda07eddb536022bca03f7fe21f6896d0106d3" exitCode=0 Dec 06 16:01:56 crc kubenswrapper[4813]: I1206 16:01:56.174184 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c96d-account-create-update-ldmnt" event={"ID":"d882403f-89d4-4bef-b2a0-7f1134e4b735","Type":"ContainerDied","Data":"f5b499d0437ecc1d705d99c428fda07eddb536022bca03f7fe21f6896d0106d3"} Dec 06 16:01:56 crc kubenswrapper[4813]: I1206 16:01:56.178105 4813 generic.go:334] "Generic (PLEG): container finished" podID="e7dd308a-25f5-4746-b850-d4a2430c6b20" containerID="97658e2d17b3cfa454f36a100545c61f2a1d47dad3dfa629978409078999cd27" exitCode=0 Dec 06 16:01:56 crc kubenswrapper[4813]: I1206 16:01:56.178250 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-2d5d-account-create-update-2pzsb" event={"ID":"e7dd308a-25f5-4746-b850-d4a2430c6b20","Type":"ContainerDied","Data":"97658e2d17b3cfa454f36a100545c61f2a1d47dad3dfa629978409078999cd27"} Dec 06 16:01:56 crc kubenswrapper[4813]: I1206 16:01:56.178328 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-2d5d-account-create-update-2pzsb" event={"ID":"e7dd308a-25f5-4746-b850-d4a2430c6b20","Type":"ContainerStarted","Data":"002a461e040d36b6f9eed06e31d1b247e72e3e68891b2bb67fa52cacd2898cef"} Dec 06 16:01:56 crc kubenswrapper[4813]: I1206 16:01:56.181431 4813 generic.go:334] "Generic (PLEG): container finished" podID="ee373829-88aa-427e-aedb-af7410b7b7a3" containerID="ec3f311f852ddce4c0ba7b5c384ed91c57f18949369b7ca524def707555f235e" exitCode=0 Dec 06 16:01:56 crc kubenswrapper[4813]: I1206 16:01:56.181590 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wt55k" event={"ID":"ee373829-88aa-427e-aedb-af7410b7b7a3","Type":"ContainerDied","Data":"ec3f311f852ddce4c0ba7b5c384ed91c57f18949369b7ca524def707555f235e"} Dec 06 16:01:56 crc kubenswrapper[4813]: I1206 16:01:56.181638 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wt55k" event={"ID":"ee373829-88aa-427e-aedb-af7410b7b7a3","Type":"ContainerStarted","Data":"727d7c8fd4edce45ba7d9bc44fc91d9ee65c68cfebe3080b44f9db1d569a0c75"} Dec 06 16:01:56 crc kubenswrapper[4813]: I1206 16:01:56.226458 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-8l7jf" Dec 06 16:01:56 crc kubenswrapper[4813]: I1206 16:01:56.330020 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-6tgdq"] Dec 06 16:01:56 crc kubenswrapper[4813]: I1206 16:01:56.330242 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-6tgdq" podUID="fb685be0-8427-43c2-ac4e-262c73714b5f" containerName="dnsmasq-dns" containerID="cri-o://13d1fa7406788ff72d6bcf5296b8907beca0f71d1196701e3d60205544a650e9" gracePeriod=10 Dec 06 16:01:56 crc kubenswrapper[4813]: I1206 16:01:56.705815 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-lwz48" Dec 06 16:01:56 crc kubenswrapper[4813]: I1206 16:01:56.753748 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5q2nc\" (UniqueName: \"kubernetes.io/projected/66c65f92-2424-4d91-9824-9a6c00bf3dbf-kube-api-access-5q2nc\") pod \"66c65f92-2424-4d91-9824-9a6c00bf3dbf\" (UID: \"66c65f92-2424-4d91-9824-9a6c00bf3dbf\") " Dec 06 16:01:56 crc kubenswrapper[4813]: I1206 16:01:56.753877 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/66c65f92-2424-4d91-9824-9a6c00bf3dbf-operator-scripts\") pod \"66c65f92-2424-4d91-9824-9a6c00bf3dbf\" (UID: \"66c65f92-2424-4d91-9824-9a6c00bf3dbf\") " Dec 06 16:01:56 crc kubenswrapper[4813]: I1206 16:01:56.754656 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66c65f92-2424-4d91-9824-9a6c00bf3dbf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "66c65f92-2424-4d91-9824-9a6c00bf3dbf" (UID: "66c65f92-2424-4d91-9824-9a6c00bf3dbf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:01:56 crc kubenswrapper[4813]: I1206 16:01:56.791512 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66c65f92-2424-4d91-9824-9a6c00bf3dbf-kube-api-access-5q2nc" (OuterVolumeSpecName: "kube-api-access-5q2nc") pod "66c65f92-2424-4d91-9824-9a6c00bf3dbf" (UID: "66c65f92-2424-4d91-9824-9a6c00bf3dbf"). InnerVolumeSpecName "kube-api-access-5q2nc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:01:56 crc kubenswrapper[4813]: I1206 16:01:56.856780 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/66c65f92-2424-4d91-9824-9a6c00bf3dbf-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:56 crc kubenswrapper[4813]: I1206 16:01:56.856816 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5q2nc\" (UniqueName: \"kubernetes.io/projected/66c65f92-2424-4d91-9824-9a6c00bf3dbf-kube-api-access-5q2nc\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:56 crc kubenswrapper[4813]: I1206 16:01:56.937681 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a434-account-create-update-5rxjv" Dec 06 16:01:56 crc kubenswrapper[4813]: I1206 16:01:56.940053 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-6tgdq" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.058819 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/854df38a-5bef-4b2c-acdc-4ca7577bc1de-operator-scripts\") pod \"854df38a-5bef-4b2c-acdc-4ca7577bc1de\" (UID: \"854df38a-5bef-4b2c-acdc-4ca7577bc1de\") " Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.058885 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb685be0-8427-43c2-ac4e-262c73714b5f-dns-svc\") pod \"fb685be0-8427-43c2-ac4e-262c73714b5f\" (UID: \"fb685be0-8427-43c2-ac4e-262c73714b5f\") " Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.058968 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5h5lk\" (UniqueName: \"kubernetes.io/projected/854df38a-5bef-4b2c-acdc-4ca7577bc1de-kube-api-access-5h5lk\") pod \"854df38a-5bef-4b2c-acdc-4ca7577bc1de\" (UID: \"854df38a-5bef-4b2c-acdc-4ca7577bc1de\") " Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.059079 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb685be0-8427-43c2-ac4e-262c73714b5f-config\") pod \"fb685be0-8427-43c2-ac4e-262c73714b5f\" (UID: \"fb685be0-8427-43c2-ac4e-262c73714b5f\") " Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.059099 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5b5x\" (UniqueName: \"kubernetes.io/projected/fb685be0-8427-43c2-ac4e-262c73714b5f-kube-api-access-d5b5x\") pod \"fb685be0-8427-43c2-ac4e-262c73714b5f\" (UID: \"fb685be0-8427-43c2-ac4e-262c73714b5f\") " Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.062430 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb685be0-8427-43c2-ac4e-262c73714b5f-kube-api-access-d5b5x" (OuterVolumeSpecName: "kube-api-access-d5b5x") pod "fb685be0-8427-43c2-ac4e-262c73714b5f" (UID: "fb685be0-8427-43c2-ac4e-262c73714b5f"). InnerVolumeSpecName "kube-api-access-d5b5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.062689 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/854df38a-5bef-4b2c-acdc-4ca7577bc1de-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "854df38a-5bef-4b2c-acdc-4ca7577bc1de" (UID: "854df38a-5bef-4b2c-acdc-4ca7577bc1de"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.073493 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/854df38a-5bef-4b2c-acdc-4ca7577bc1de-kube-api-access-5h5lk" (OuterVolumeSpecName: "kube-api-access-5h5lk") pod "854df38a-5bef-4b2c-acdc-4ca7577bc1de" (UID: "854df38a-5bef-4b2c-acdc-4ca7577bc1de"). InnerVolumeSpecName "kube-api-access-5h5lk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.124873 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb685be0-8427-43c2-ac4e-262c73714b5f-config" (OuterVolumeSpecName: "config") pod "fb685be0-8427-43c2-ac4e-262c73714b5f" (UID: "fb685be0-8427-43c2-ac4e-262c73714b5f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.136815 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb685be0-8427-43c2-ac4e-262c73714b5f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fb685be0-8427-43c2-ac4e-262c73714b5f" (UID: "fb685be0-8427-43c2-ac4e-262c73714b5f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.161298 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb685be0-8427-43c2-ac4e-262c73714b5f-config\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.161344 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5b5x\" (UniqueName: \"kubernetes.io/projected/fb685be0-8427-43c2-ac4e-262c73714b5f-kube-api-access-d5b5x\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.161358 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/854df38a-5bef-4b2c-acdc-4ca7577bc1de-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.161373 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb685be0-8427-43c2-ac4e-262c73714b5f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.161385 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5h5lk\" (UniqueName: \"kubernetes.io/projected/854df38a-5bef-4b2c-acdc-4ca7577bc1de-kube-api-access-5h5lk\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.231727 4813 generic.go:334] "Generic (PLEG): container finished" podID="fb685be0-8427-43c2-ac4e-262c73714b5f" containerID="13d1fa7406788ff72d6bcf5296b8907beca0f71d1196701e3d60205544a650e9" exitCode=0 Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.231797 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-6tgdq" event={"ID":"fb685be0-8427-43c2-ac4e-262c73714b5f","Type":"ContainerDied","Data":"13d1fa7406788ff72d6bcf5296b8907beca0f71d1196701e3d60205544a650e9"} Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.231824 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-6tgdq" event={"ID":"fb685be0-8427-43c2-ac4e-262c73714b5f","Type":"ContainerDied","Data":"6016788dc03ab2801e4b984ea374765bae15fd5d2d6c7fdb1b318c0c3b64ac99"} Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.231845 4813 scope.go:117] "RemoveContainer" containerID="13d1fa7406788ff72d6bcf5296b8907beca0f71d1196701e3d60205544a650e9" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.231996 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-6tgdq" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.243212 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a434-account-create-update-5rxjv" event={"ID":"854df38a-5bef-4b2c-acdc-4ca7577bc1de","Type":"ContainerDied","Data":"d0487c8f9062dd074487c1690aad9e1450b105b09a47f33a3ea823e92727d6ce"} Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.243252 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0487c8f9062dd074487c1690aad9e1450b105b09a47f33a3ea823e92727d6ce" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.243326 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a434-account-create-update-5rxjv" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.257126 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-lwz48" event={"ID":"66c65f92-2424-4d91-9824-9a6c00bf3dbf","Type":"ContainerDied","Data":"e8c6183bc681ff680bdbd9e98d28ebcd4587072c70740e476d7cd61fe6df1583"} Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.257171 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8c6183bc681ff680bdbd9e98d28ebcd4587072c70740e476d7cd61fe6df1583" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.257196 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-lwz48" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.270510 4813 scope.go:117] "RemoveContainer" containerID="3165b9a67a884f501a9e9e0fba148fb0f0151d101da7a4e0efdbdf19fac09b04" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.272570 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-6tgdq"] Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.289093 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-6tgdq"] Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.302882 4813 scope.go:117] "RemoveContainer" containerID="13d1fa7406788ff72d6bcf5296b8907beca0f71d1196701e3d60205544a650e9" Dec 06 16:01:57 crc kubenswrapper[4813]: E1206 16:01:57.303725 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13d1fa7406788ff72d6bcf5296b8907beca0f71d1196701e3d60205544a650e9\": container with ID starting with 13d1fa7406788ff72d6bcf5296b8907beca0f71d1196701e3d60205544a650e9 not found: ID does not exist" containerID="13d1fa7406788ff72d6bcf5296b8907beca0f71d1196701e3d60205544a650e9" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.303769 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13d1fa7406788ff72d6bcf5296b8907beca0f71d1196701e3d60205544a650e9"} err="failed to get container status \"13d1fa7406788ff72d6bcf5296b8907beca0f71d1196701e3d60205544a650e9\": rpc error: code = NotFound desc = could not find container \"13d1fa7406788ff72d6bcf5296b8907beca0f71d1196701e3d60205544a650e9\": container with ID starting with 13d1fa7406788ff72d6bcf5296b8907beca0f71d1196701e3d60205544a650e9 not found: ID does not exist" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.303796 4813 scope.go:117] "RemoveContainer" containerID="3165b9a67a884f501a9e9e0fba148fb0f0151d101da7a4e0efdbdf19fac09b04" Dec 06 16:01:57 crc kubenswrapper[4813]: E1206 16:01:57.304008 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3165b9a67a884f501a9e9e0fba148fb0f0151d101da7a4e0efdbdf19fac09b04\": container with ID starting with 3165b9a67a884f501a9e9e0fba148fb0f0151d101da7a4e0efdbdf19fac09b04 not found: ID does not exist" containerID="3165b9a67a884f501a9e9e0fba148fb0f0151d101da7a4e0efdbdf19fac09b04" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.304029 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3165b9a67a884f501a9e9e0fba148fb0f0151d101da7a4e0efdbdf19fac09b04"} err="failed to get container status \"3165b9a67a884f501a9e9e0fba148fb0f0151d101da7a4e0efdbdf19fac09b04\": rpc error: code = NotFound desc = could not find container \"3165b9a67a884f501a9e9e0fba148fb0f0151d101da7a4e0efdbdf19fac09b04\": container with ID starting with 3165b9a67a884f501a9e9e0fba148fb0f0151d101da7a4e0efdbdf19fac09b04 not found: ID does not exist" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.609277 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-2d5d-account-create-update-2pzsb" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.673337 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7dd308a-25f5-4746-b850-d4a2430c6b20-operator-scripts\") pod \"e7dd308a-25f5-4746-b850-d4a2430c6b20\" (UID: \"e7dd308a-25f5-4746-b850-d4a2430c6b20\") " Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.673554 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wb7nl\" (UniqueName: \"kubernetes.io/projected/e7dd308a-25f5-4746-b850-d4a2430c6b20-kube-api-access-wb7nl\") pod \"e7dd308a-25f5-4746-b850-d4a2430c6b20\" (UID: \"e7dd308a-25f5-4746-b850-d4a2430c6b20\") " Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.674207 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7dd308a-25f5-4746-b850-d4a2430c6b20-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e7dd308a-25f5-4746-b850-d4a2430c6b20" (UID: "e7dd308a-25f5-4746-b850-d4a2430c6b20"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.679334 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7dd308a-25f5-4746-b850-d4a2430c6b20-kube-api-access-wb7nl" (OuterVolumeSpecName: "kube-api-access-wb7nl") pod "e7dd308a-25f5-4746-b850-d4a2430c6b20" (UID: "e7dd308a-25f5-4746-b850-d4a2430c6b20"). InnerVolumeSpecName "kube-api-access-wb7nl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.777149 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wb7nl\" (UniqueName: \"kubernetes.io/projected/e7dd308a-25f5-4746-b850-d4a2430c6b20-kube-api-access-wb7nl\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.777194 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7dd308a-25f5-4746-b850-d4a2430c6b20-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.847267 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-qcgfq" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.852078 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wt55k" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.873067 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c96d-account-create-update-ldmnt" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.878156 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6805af73-82e1-4e15-b867-d02abc99d5ee-operator-scripts\") pod \"6805af73-82e1-4e15-b867-d02abc99d5ee\" (UID: \"6805af73-82e1-4e15-b867-d02abc99d5ee\") " Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.878674 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6805af73-82e1-4e15-b867-d02abc99d5ee-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6805af73-82e1-4e15-b867-d02abc99d5ee" (UID: "6805af73-82e1-4e15-b867-d02abc99d5ee"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.980445 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9tnzt\" (UniqueName: \"kubernetes.io/projected/d882403f-89d4-4bef-b2a0-7f1134e4b735-kube-api-access-9tnzt\") pod \"d882403f-89d4-4bef-b2a0-7f1134e4b735\" (UID: \"d882403f-89d4-4bef-b2a0-7f1134e4b735\") " Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.980523 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwzqm\" (UniqueName: \"kubernetes.io/projected/ee373829-88aa-427e-aedb-af7410b7b7a3-kube-api-access-zwzqm\") pod \"ee373829-88aa-427e-aedb-af7410b7b7a3\" (UID: \"ee373829-88aa-427e-aedb-af7410b7b7a3\") " Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.980578 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee373829-88aa-427e-aedb-af7410b7b7a3-operator-scripts\") pod \"ee373829-88aa-427e-aedb-af7410b7b7a3\" (UID: \"ee373829-88aa-427e-aedb-af7410b7b7a3\") " Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.980609 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d882403f-89d4-4bef-b2a0-7f1134e4b735-operator-scripts\") pod \"d882403f-89d4-4bef-b2a0-7f1134e4b735\" (UID: \"d882403f-89d4-4bef-b2a0-7f1134e4b735\") " Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.980633 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfz7l\" (UniqueName: \"kubernetes.io/projected/6805af73-82e1-4e15-b867-d02abc99d5ee-kube-api-access-qfz7l\") pod \"6805af73-82e1-4e15-b867-d02abc99d5ee\" (UID: \"6805af73-82e1-4e15-b867-d02abc99d5ee\") " Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.981003 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6805af73-82e1-4e15-b867-d02abc99d5ee-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.982208 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d882403f-89d4-4bef-b2a0-7f1134e4b735-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d882403f-89d4-4bef-b2a0-7f1134e4b735" (UID: "d882403f-89d4-4bef-b2a0-7f1134e4b735"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.982607 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee373829-88aa-427e-aedb-af7410b7b7a3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ee373829-88aa-427e-aedb-af7410b7b7a3" (UID: "ee373829-88aa-427e-aedb-af7410b7b7a3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.987533 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee373829-88aa-427e-aedb-af7410b7b7a3-kube-api-access-zwzqm" (OuterVolumeSpecName: "kube-api-access-zwzqm") pod "ee373829-88aa-427e-aedb-af7410b7b7a3" (UID: "ee373829-88aa-427e-aedb-af7410b7b7a3"). InnerVolumeSpecName "kube-api-access-zwzqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.987650 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6805af73-82e1-4e15-b867-d02abc99d5ee-kube-api-access-qfz7l" (OuterVolumeSpecName: "kube-api-access-qfz7l") pod "6805af73-82e1-4e15-b867-d02abc99d5ee" (UID: "6805af73-82e1-4e15-b867-d02abc99d5ee"). InnerVolumeSpecName "kube-api-access-qfz7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:01:57 crc kubenswrapper[4813]: I1206 16:01:57.989432 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d882403f-89d4-4bef-b2a0-7f1134e4b735-kube-api-access-9tnzt" (OuterVolumeSpecName: "kube-api-access-9tnzt") pod "d882403f-89d4-4bef-b2a0-7f1134e4b735" (UID: "d882403f-89d4-4bef-b2a0-7f1134e4b735"). InnerVolumeSpecName "kube-api-access-9tnzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:01:58 crc kubenswrapper[4813]: I1206 16:01:58.082637 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9tnzt\" (UniqueName: \"kubernetes.io/projected/d882403f-89d4-4bef-b2a0-7f1134e4b735-kube-api-access-9tnzt\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:58 crc kubenswrapper[4813]: I1206 16:01:58.082682 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwzqm\" (UniqueName: \"kubernetes.io/projected/ee373829-88aa-427e-aedb-af7410b7b7a3-kube-api-access-zwzqm\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:58 crc kubenswrapper[4813]: I1206 16:01:58.082701 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee373829-88aa-427e-aedb-af7410b7b7a3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:58 crc kubenswrapper[4813]: I1206 16:01:58.082718 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d882403f-89d4-4bef-b2a0-7f1134e4b735-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:58 crc kubenswrapper[4813]: I1206 16:01:58.082735 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfz7l\" (UniqueName: \"kubernetes.io/projected/6805af73-82e1-4e15-b867-d02abc99d5ee-kube-api-access-qfz7l\") on node \"crc\" DevicePath \"\"" Dec 06 16:01:58 crc kubenswrapper[4813]: I1206 16:01:58.269182 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-qcgfq" event={"ID":"6805af73-82e1-4e15-b867-d02abc99d5ee","Type":"ContainerDied","Data":"1434cc76fe9a6998b70c27aa9043e85711adeccfeef1cb0a4f740e5d17567576"} Dec 06 16:01:58 crc kubenswrapper[4813]: I1206 16:01:58.269212 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-qcgfq" Dec 06 16:01:58 crc kubenswrapper[4813]: I1206 16:01:58.269239 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1434cc76fe9a6998b70c27aa9043e85711adeccfeef1cb0a4f740e5d17567576" Dec 06 16:01:58 crc kubenswrapper[4813]: I1206 16:01:58.274489 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c96d-account-create-update-ldmnt" event={"ID":"d882403f-89d4-4bef-b2a0-7f1134e4b735","Type":"ContainerDied","Data":"e0b38d8afd058fec9102e0f80619790089b4597cc928a5c12e048fb6b9723417"} Dec 06 16:01:58 crc kubenswrapper[4813]: I1206 16:01:58.274525 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c96d-account-create-update-ldmnt" Dec 06 16:01:58 crc kubenswrapper[4813]: I1206 16:01:58.274553 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0b38d8afd058fec9102e0f80619790089b4597cc928a5c12e048fb6b9723417" Dec 06 16:01:58 crc kubenswrapper[4813]: I1206 16:01:58.278092 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-2d5d-account-create-update-2pzsb" event={"ID":"e7dd308a-25f5-4746-b850-d4a2430c6b20","Type":"ContainerDied","Data":"002a461e040d36b6f9eed06e31d1b247e72e3e68891b2bb67fa52cacd2898cef"} Dec 06 16:01:58 crc kubenswrapper[4813]: I1206 16:01:58.278128 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="002a461e040d36b6f9eed06e31d1b247e72e3e68891b2bb67fa52cacd2898cef" Dec 06 16:01:58 crc kubenswrapper[4813]: I1206 16:01:58.278199 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-2d5d-account-create-update-2pzsb" Dec 06 16:01:58 crc kubenswrapper[4813]: I1206 16:01:58.284723 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wt55k" event={"ID":"ee373829-88aa-427e-aedb-af7410b7b7a3","Type":"ContainerDied","Data":"727d7c8fd4edce45ba7d9bc44fc91d9ee65c68cfebe3080b44f9db1d569a0c75"} Dec 06 16:01:58 crc kubenswrapper[4813]: I1206 16:01:58.284760 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="727d7c8fd4edce45ba7d9bc44fc91d9ee65c68cfebe3080b44f9db1d569a0c75" Dec 06 16:01:58 crc kubenswrapper[4813]: I1206 16:01:58.284862 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wt55k" Dec 06 16:01:58 crc kubenswrapper[4813]: I1206 16:01:58.505396 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb685be0-8427-43c2-ac4e-262c73714b5f" path="/var/lib/kubelet/pods/fb685be0-8427-43c2-ac4e-262c73714b5f/volumes" Dec 06 16:01:59 crc kubenswrapper[4813]: I1206 16:01:59.852426 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-tzd8d"] Dec 06 16:01:59 crc kubenswrapper[4813]: E1206 16:01:59.852985 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7dd308a-25f5-4746-b850-d4a2430c6b20" containerName="mariadb-account-create-update" Dec 06 16:01:59 crc kubenswrapper[4813]: I1206 16:01:59.852996 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7dd308a-25f5-4746-b850-d4a2430c6b20" containerName="mariadb-account-create-update" Dec 06 16:01:59 crc kubenswrapper[4813]: E1206 16:01:59.853007 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6805af73-82e1-4e15-b867-d02abc99d5ee" containerName="mariadb-database-create" Dec 06 16:01:59 crc kubenswrapper[4813]: I1206 16:01:59.853013 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="6805af73-82e1-4e15-b867-d02abc99d5ee" containerName="mariadb-database-create" Dec 06 16:01:59 crc kubenswrapper[4813]: E1206 16:01:59.853031 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee373829-88aa-427e-aedb-af7410b7b7a3" containerName="mariadb-database-create" Dec 06 16:01:59 crc kubenswrapper[4813]: I1206 16:01:59.853036 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee373829-88aa-427e-aedb-af7410b7b7a3" containerName="mariadb-database-create" Dec 06 16:01:59 crc kubenswrapper[4813]: E1206 16:01:59.853045 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d882403f-89d4-4bef-b2a0-7f1134e4b735" containerName="mariadb-account-create-update" Dec 06 16:01:59 crc kubenswrapper[4813]: I1206 16:01:59.853051 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d882403f-89d4-4bef-b2a0-7f1134e4b735" containerName="mariadb-account-create-update" Dec 06 16:01:59 crc kubenswrapper[4813]: E1206 16:01:59.853066 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb685be0-8427-43c2-ac4e-262c73714b5f" containerName="init" Dec 06 16:01:59 crc kubenswrapper[4813]: I1206 16:01:59.853073 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb685be0-8427-43c2-ac4e-262c73714b5f" containerName="init" Dec 06 16:01:59 crc kubenswrapper[4813]: E1206 16:01:59.853093 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb685be0-8427-43c2-ac4e-262c73714b5f" containerName="dnsmasq-dns" Dec 06 16:01:59 crc kubenswrapper[4813]: I1206 16:01:59.853099 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb685be0-8427-43c2-ac4e-262c73714b5f" containerName="dnsmasq-dns" Dec 06 16:01:59 crc kubenswrapper[4813]: E1206 16:01:59.853109 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="854df38a-5bef-4b2c-acdc-4ca7577bc1de" containerName="mariadb-account-create-update" Dec 06 16:01:59 crc kubenswrapper[4813]: I1206 16:01:59.853115 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="854df38a-5bef-4b2c-acdc-4ca7577bc1de" containerName="mariadb-account-create-update" Dec 06 16:01:59 crc kubenswrapper[4813]: E1206 16:01:59.853129 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66c65f92-2424-4d91-9824-9a6c00bf3dbf" containerName="mariadb-database-create" Dec 06 16:01:59 crc kubenswrapper[4813]: I1206 16:01:59.853134 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="66c65f92-2424-4d91-9824-9a6c00bf3dbf" containerName="mariadb-database-create" Dec 06 16:01:59 crc kubenswrapper[4813]: I1206 16:01:59.853284 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="854df38a-5bef-4b2c-acdc-4ca7577bc1de" containerName="mariadb-account-create-update" Dec 06 16:01:59 crc kubenswrapper[4813]: I1206 16:01:59.853295 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d882403f-89d4-4bef-b2a0-7f1134e4b735" containerName="mariadb-account-create-update" Dec 06 16:01:59 crc kubenswrapper[4813]: I1206 16:01:59.853306 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="6805af73-82e1-4e15-b867-d02abc99d5ee" containerName="mariadb-database-create" Dec 06 16:01:59 crc kubenswrapper[4813]: I1206 16:01:59.853314 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee373829-88aa-427e-aedb-af7410b7b7a3" containerName="mariadb-database-create" Dec 06 16:01:59 crc kubenswrapper[4813]: I1206 16:01:59.853326 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="66c65f92-2424-4d91-9824-9a6c00bf3dbf" containerName="mariadb-database-create" Dec 06 16:01:59 crc kubenswrapper[4813]: I1206 16:01:59.853336 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb685be0-8427-43c2-ac4e-262c73714b5f" containerName="dnsmasq-dns" Dec 06 16:01:59 crc kubenswrapper[4813]: I1206 16:01:59.853345 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7dd308a-25f5-4746-b850-d4a2430c6b20" containerName="mariadb-account-create-update" Dec 06 16:01:59 crc kubenswrapper[4813]: I1206 16:01:59.853801 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tzd8d" Dec 06 16:01:59 crc kubenswrapper[4813]: I1206 16:01:59.856447 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-sxdvk" Dec 06 16:01:59 crc kubenswrapper[4813]: I1206 16:01:59.856802 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 06 16:01:59 crc kubenswrapper[4813]: I1206 16:01:59.864844 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-tzd8d"] Dec 06 16:01:59 crc kubenswrapper[4813]: I1206 16:01:59.930865 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5xhm\" (UniqueName: \"kubernetes.io/projected/27b4a8bf-5c37-4936-8ced-6a1d02249dd1-kube-api-access-f5xhm\") pod \"glance-db-sync-tzd8d\" (UID: \"27b4a8bf-5c37-4936-8ced-6a1d02249dd1\") " pod="openstack/glance-db-sync-tzd8d" Dec 06 16:01:59 crc kubenswrapper[4813]: I1206 16:01:59.930931 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27b4a8bf-5c37-4936-8ced-6a1d02249dd1-config-data\") pod \"glance-db-sync-tzd8d\" (UID: \"27b4a8bf-5c37-4936-8ced-6a1d02249dd1\") " pod="openstack/glance-db-sync-tzd8d" Dec 06 16:01:59 crc kubenswrapper[4813]: I1206 16:01:59.930964 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/27b4a8bf-5c37-4936-8ced-6a1d02249dd1-db-sync-config-data\") pod \"glance-db-sync-tzd8d\" (UID: \"27b4a8bf-5c37-4936-8ced-6a1d02249dd1\") " pod="openstack/glance-db-sync-tzd8d" Dec 06 16:01:59 crc kubenswrapper[4813]: I1206 16:01:59.931071 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27b4a8bf-5c37-4936-8ced-6a1d02249dd1-combined-ca-bundle\") pod \"glance-db-sync-tzd8d\" (UID: \"27b4a8bf-5c37-4936-8ced-6a1d02249dd1\") " pod="openstack/glance-db-sync-tzd8d" Dec 06 16:02:00 crc kubenswrapper[4813]: I1206 16:02:00.032477 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27b4a8bf-5c37-4936-8ced-6a1d02249dd1-combined-ca-bundle\") pod \"glance-db-sync-tzd8d\" (UID: \"27b4a8bf-5c37-4936-8ced-6a1d02249dd1\") " pod="openstack/glance-db-sync-tzd8d" Dec 06 16:02:00 crc kubenswrapper[4813]: I1206 16:02:00.032564 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5xhm\" (UniqueName: \"kubernetes.io/projected/27b4a8bf-5c37-4936-8ced-6a1d02249dd1-kube-api-access-f5xhm\") pod \"glance-db-sync-tzd8d\" (UID: \"27b4a8bf-5c37-4936-8ced-6a1d02249dd1\") " pod="openstack/glance-db-sync-tzd8d" Dec 06 16:02:00 crc kubenswrapper[4813]: I1206 16:02:00.032601 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27b4a8bf-5c37-4936-8ced-6a1d02249dd1-config-data\") pod \"glance-db-sync-tzd8d\" (UID: \"27b4a8bf-5c37-4936-8ced-6a1d02249dd1\") " pod="openstack/glance-db-sync-tzd8d" Dec 06 16:02:00 crc kubenswrapper[4813]: I1206 16:02:00.032630 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/27b4a8bf-5c37-4936-8ced-6a1d02249dd1-db-sync-config-data\") pod \"glance-db-sync-tzd8d\" (UID: \"27b4a8bf-5c37-4936-8ced-6a1d02249dd1\") " pod="openstack/glance-db-sync-tzd8d" Dec 06 16:02:00 crc kubenswrapper[4813]: I1206 16:02:00.037826 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/27b4a8bf-5c37-4936-8ced-6a1d02249dd1-db-sync-config-data\") pod \"glance-db-sync-tzd8d\" (UID: \"27b4a8bf-5c37-4936-8ced-6a1d02249dd1\") " pod="openstack/glance-db-sync-tzd8d" Dec 06 16:02:00 crc kubenswrapper[4813]: I1206 16:02:00.038034 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27b4a8bf-5c37-4936-8ced-6a1d02249dd1-combined-ca-bundle\") pod \"glance-db-sync-tzd8d\" (UID: \"27b4a8bf-5c37-4936-8ced-6a1d02249dd1\") " pod="openstack/glance-db-sync-tzd8d" Dec 06 16:02:00 crc kubenswrapper[4813]: I1206 16:02:00.041651 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27b4a8bf-5c37-4936-8ced-6a1d02249dd1-config-data\") pod \"glance-db-sync-tzd8d\" (UID: \"27b4a8bf-5c37-4936-8ced-6a1d02249dd1\") " pod="openstack/glance-db-sync-tzd8d" Dec 06 16:02:00 crc kubenswrapper[4813]: I1206 16:02:00.061348 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5xhm\" (UniqueName: \"kubernetes.io/projected/27b4a8bf-5c37-4936-8ced-6a1d02249dd1-kube-api-access-f5xhm\") pod \"glance-db-sync-tzd8d\" (UID: \"27b4a8bf-5c37-4936-8ced-6a1d02249dd1\") " pod="openstack/glance-db-sync-tzd8d" Dec 06 16:02:00 crc kubenswrapper[4813]: I1206 16:02:00.172339 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tzd8d" Dec 06 16:02:00 crc kubenswrapper[4813]: I1206 16:02:00.364206 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-bg6cb" podUID="ba6825c1-2137-46d6-a695-6bdd53a13ec4" containerName="ovn-controller" probeResult="failure" output=< Dec 06 16:02:00 crc kubenswrapper[4813]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 06 16:02:00 crc kubenswrapper[4813]: > Dec 06 16:02:00 crc kubenswrapper[4813]: I1206 16:02:00.647794 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 06 16:02:00 crc kubenswrapper[4813]: I1206 16:02:00.776871 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-tzd8d"] Dec 06 16:02:00 crc kubenswrapper[4813]: W1206 16:02:00.777962 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27b4a8bf_5c37_4936_8ced_6a1d02249dd1.slice/crio-291932d28876c2271569b42c8c1d40cf9cf7da40dc6dd77155b98e1840d90b48 WatchSource:0}: Error finding container 291932d28876c2271569b42c8c1d40cf9cf7da40dc6dd77155b98e1840d90b48: Status 404 returned error can't find the container with id 291932d28876c2271569b42c8c1d40cf9cf7da40dc6dd77155b98e1840d90b48 Dec 06 16:02:01 crc kubenswrapper[4813]: I1206 16:02:01.317224 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tzd8d" event={"ID":"27b4a8bf-5c37-4936-8ced-6a1d02249dd1","Type":"ContainerStarted","Data":"291932d28876c2271569b42c8c1d40cf9cf7da40dc6dd77155b98e1840d90b48"} Dec 06 16:02:02 crc kubenswrapper[4813]: I1206 16:02:02.326089 4813 generic.go:334] "Generic (PLEG): container finished" podID="f67c4c64-2c67-46bc-af93-6ca6ac4f4814" containerID="bd9877744dc4c39dbe47b728f189edbcd112064748429eec348a62b3b78bc0e9" exitCode=0 Dec 06 16:02:02 crc kubenswrapper[4813]: I1206 16:02:02.326137 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-7pplm" event={"ID":"f67c4c64-2c67-46bc-af93-6ca6ac4f4814","Type":"ContainerDied","Data":"bd9877744dc4c39dbe47b728f189edbcd112064748429eec348a62b3b78bc0e9"} Dec 06 16:02:02 crc kubenswrapper[4813]: I1206 16:02:02.900501 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4f571761-c8de-485f-89de-514fc0a66a55-etc-swift\") pod \"swift-storage-0\" (UID: \"4f571761-c8de-485f-89de-514fc0a66a55\") " pod="openstack/swift-storage-0" Dec 06 16:02:02 crc kubenswrapper[4813]: I1206 16:02:02.909962 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4f571761-c8de-485f-89de-514fc0a66a55-etc-swift\") pod \"swift-storage-0\" (UID: \"4f571761-c8de-485f-89de-514fc0a66a55\") " pod="openstack/swift-storage-0" Dec 06 16:02:02 crc kubenswrapper[4813]: I1206 16:02:02.956980 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 06 16:02:03 crc kubenswrapper[4813]: I1206 16:02:03.341758 4813 generic.go:334] "Generic (PLEG): container finished" podID="2172dfb5-c0a6-4525-a326-dd9736eb5233" containerID="2a1916b4fe499237d82c88926954303ec3c62edb0721baeedf52bceebb68441b" exitCode=0 Dec 06 16:02:03 crc kubenswrapper[4813]: I1206 16:02:03.342142 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2172dfb5-c0a6-4525-a326-dd9736eb5233","Type":"ContainerDied","Data":"2a1916b4fe499237d82c88926954303ec3c62edb0721baeedf52bceebb68441b"} Dec 06 16:02:03 crc kubenswrapper[4813]: I1206 16:02:03.346630 4813 generic.go:334] "Generic (PLEG): container finished" podID="f03c3251-5e1c-4cb2-aa3f-ebbef28acb19" containerID="d2852b625a8e7a7b46550b6c2cafd975948130b8c072519d3613885a39ad6c8f" exitCode=0 Dec 06 16:02:03 crc kubenswrapper[4813]: I1206 16:02:03.346675 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19","Type":"ContainerDied","Data":"d2852b625a8e7a7b46550b6c2cafd975948130b8c072519d3613885a39ad6c8f"} Dec 06 16:02:03 crc kubenswrapper[4813]: I1206 16:02:03.552077 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 06 16:02:03 crc kubenswrapper[4813]: I1206 16:02:03.640755 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-7pplm" Dec 06 16:02:03 crc kubenswrapper[4813]: I1206 16:02:03.712444 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-ring-data-devices\") pod \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\" (UID: \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\") " Dec 06 16:02:03 crc kubenswrapper[4813]: I1206 16:02:03.712482 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9xg8\" (UniqueName: \"kubernetes.io/projected/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-kube-api-access-h9xg8\") pod \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\" (UID: \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\") " Dec 06 16:02:03 crc kubenswrapper[4813]: I1206 16:02:03.712505 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-combined-ca-bundle\") pod \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\" (UID: \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\") " Dec 06 16:02:03 crc kubenswrapper[4813]: I1206 16:02:03.712573 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-scripts\") pod \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\" (UID: \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\") " Dec 06 16:02:03 crc kubenswrapper[4813]: I1206 16:02:03.712647 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-etc-swift\") pod \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\" (UID: \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\") " Dec 06 16:02:03 crc kubenswrapper[4813]: I1206 16:02:03.712709 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-dispersionconf\") pod \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\" (UID: \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\") " Dec 06 16:02:03 crc kubenswrapper[4813]: I1206 16:02:03.712726 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-swiftconf\") pod \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\" (UID: \"f67c4c64-2c67-46bc-af93-6ca6ac4f4814\") " Dec 06 16:02:03 crc kubenswrapper[4813]: I1206 16:02:03.713241 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "f67c4c64-2c67-46bc-af93-6ca6ac4f4814" (UID: "f67c4c64-2c67-46bc-af93-6ca6ac4f4814"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:02:03 crc kubenswrapper[4813]: I1206 16:02:03.714202 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "f67c4c64-2c67-46bc-af93-6ca6ac4f4814" (UID: "f67c4c64-2c67-46bc-af93-6ca6ac4f4814"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:02:03 crc kubenswrapper[4813]: I1206 16:02:03.722368 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-kube-api-access-h9xg8" (OuterVolumeSpecName: "kube-api-access-h9xg8") pod "f67c4c64-2c67-46bc-af93-6ca6ac4f4814" (UID: "f67c4c64-2c67-46bc-af93-6ca6ac4f4814"). InnerVolumeSpecName "kube-api-access-h9xg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:02:03 crc kubenswrapper[4813]: I1206 16:02:03.723410 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "f67c4c64-2c67-46bc-af93-6ca6ac4f4814" (UID: "f67c4c64-2c67-46bc-af93-6ca6ac4f4814"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:02:03 crc kubenswrapper[4813]: I1206 16:02:03.735304 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f67c4c64-2c67-46bc-af93-6ca6ac4f4814" (UID: "f67c4c64-2c67-46bc-af93-6ca6ac4f4814"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:02:03 crc kubenswrapper[4813]: I1206 16:02:03.737621 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-scripts" (OuterVolumeSpecName: "scripts") pod "f67c4c64-2c67-46bc-af93-6ca6ac4f4814" (UID: "f67c4c64-2c67-46bc-af93-6ca6ac4f4814"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:02:03 crc kubenswrapper[4813]: I1206 16:02:03.740547 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "f67c4c64-2c67-46bc-af93-6ca6ac4f4814" (UID: "f67c4c64-2c67-46bc-af93-6ca6ac4f4814"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:02:03 crc kubenswrapper[4813]: I1206 16:02:03.814756 4813 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:03 crc kubenswrapper[4813]: I1206 16:02:03.814787 4813 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:03 crc kubenswrapper[4813]: I1206 16:02:03.814797 4813 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:03 crc kubenswrapper[4813]: I1206 16:02:03.814805 4813 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:03 crc kubenswrapper[4813]: I1206 16:02:03.814814 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9xg8\" (UniqueName: \"kubernetes.io/projected/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-kube-api-access-h9xg8\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:03 crc kubenswrapper[4813]: I1206 16:02:03.814823 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:03 crc kubenswrapper[4813]: I1206 16:02:03.814832 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f67c4c64-2c67-46bc-af93-6ca6ac4f4814-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:04 crc kubenswrapper[4813]: I1206 16:02:04.356611 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-7pplm" event={"ID":"f67c4c64-2c67-46bc-af93-6ca6ac4f4814","Type":"ContainerDied","Data":"feae26864325a76ec28c1c9f07e5b42eb3ad928af4fb1104f03b55bdcbb5682e"} Dec 06 16:02:04 crc kubenswrapper[4813]: I1206 16:02:04.356858 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="feae26864325a76ec28c1c9f07e5b42eb3ad928af4fb1104f03b55bdcbb5682e" Dec 06 16:02:04 crc kubenswrapper[4813]: I1206 16:02:04.356925 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-7pplm" Dec 06 16:02:04 crc kubenswrapper[4813]: I1206 16:02:04.361574 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2172dfb5-c0a6-4525-a326-dd9736eb5233","Type":"ContainerStarted","Data":"43a1c843efdf2356e06677d802b023e646e54b4c679dfa4d791704c633a487bd"} Dec 06 16:02:04 crc kubenswrapper[4813]: I1206 16:02:04.362571 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:02:04 crc kubenswrapper[4813]: I1206 16:02:04.371374 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f571761-c8de-485f-89de-514fc0a66a55","Type":"ContainerStarted","Data":"adf9d458411643ddbebfd75cd8998fbb05e5f757b2a67eedc35b549af8c18adb"} Dec 06 16:02:04 crc kubenswrapper[4813]: I1206 16:02:04.378684 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19","Type":"ContainerStarted","Data":"8e69c41db3d3d7e419c2bdde0dabb053b2cf3846832653472e595e340e23a8a7"} Dec 06 16:02:04 crc kubenswrapper[4813]: I1206 16:02:04.378986 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 06 16:02:04 crc kubenswrapper[4813]: I1206 16:02:04.412499 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=55.388377874 podStartE2EDuration="1m6.412483649s" podCreationTimestamp="2025-12-06 16:00:58 +0000 UTC" firstStartedPulling="2025-12-06 16:01:17.688018859 +0000 UTC m=+917.578898425" lastFinishedPulling="2025-12-06 16:01:28.712124624 +0000 UTC m=+928.603004200" observedRunningTime="2025-12-06 16:02:04.387844507 +0000 UTC m=+964.278724083" watchObservedRunningTime="2025-12-06 16:02:04.412483649 +0000 UTC m=+964.303363225" Dec 06 16:02:04 crc kubenswrapper[4813]: I1206 16:02:04.418332 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=54.767193122 podStartE2EDuration="1m5.418317394s" podCreationTimestamp="2025-12-06 16:00:59 +0000 UTC" firstStartedPulling="2025-12-06 16:01:17.741543736 +0000 UTC m=+917.632423312" lastFinishedPulling="2025-12-06 16:01:28.392668008 +0000 UTC m=+928.283547584" observedRunningTime="2025-12-06 16:02:04.412379027 +0000 UTC m=+964.303258603" watchObservedRunningTime="2025-12-06 16:02:04.418317394 +0000 UTC m=+964.309196970" Dec 06 16:02:05 crc kubenswrapper[4813]: I1206 16:02:05.346691 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-bg6cb" podUID="ba6825c1-2137-46d6-a695-6bdd53a13ec4" containerName="ovn-controller" probeResult="failure" output=< Dec 06 16:02:05 crc kubenswrapper[4813]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 06 16:02:05 crc kubenswrapper[4813]: > Dec 06 16:02:05 crc kubenswrapper[4813]: I1206 16:02:05.387786 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f571761-c8de-485f-89de-514fc0a66a55","Type":"ContainerStarted","Data":"a4c9f0cd9dcce6d6fe7a933f8309a7818f77321d0f63191289a3e2a1e8fa30cb"} Dec 06 16:02:05 crc kubenswrapper[4813]: I1206 16:02:05.388102 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f571761-c8de-485f-89de-514fc0a66a55","Type":"ContainerStarted","Data":"45a55ad3bbae3ac7c26ebad659ec6c44a36b0bbdcddcaecad6c0ccf3bfee76c0"} Dec 06 16:02:05 crc kubenswrapper[4813]: I1206 16:02:05.447743 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-mk4nr" Dec 06 16:02:06 crc kubenswrapper[4813]: I1206 16:02:06.398403 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f571761-c8de-485f-89de-514fc0a66a55","Type":"ContainerStarted","Data":"d4a410fe345785d7b6c9cb083e696f913342b20e63618ec004184355b60e470a"} Dec 06 16:02:06 crc kubenswrapper[4813]: I1206 16:02:06.398646 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f571761-c8de-485f-89de-514fc0a66a55","Type":"ContainerStarted","Data":"bb94e24b1579d0692b8e1c7aaa2a04e466b42f0dfa7f46c580cbbd01fb3848ec"} Dec 06 16:02:07 crc kubenswrapper[4813]: I1206 16:02:07.418376 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f571761-c8de-485f-89de-514fc0a66a55","Type":"ContainerStarted","Data":"5d0c546a0b76035e91adf167dc7ab2783ceab64b6ff779fd59ff7042c6ccea30"} Dec 06 16:02:07 crc kubenswrapper[4813]: I1206 16:02:07.418442 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f571761-c8de-485f-89de-514fc0a66a55","Type":"ContainerStarted","Data":"91c2146bd07763fd363abd6d08d8a2d461162518ed30ceba22b67367b7b7b13e"} Dec 06 16:02:10 crc kubenswrapper[4813]: I1206 16:02:10.323723 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-bg6cb" podUID="ba6825c1-2137-46d6-a695-6bdd53a13ec4" containerName="ovn-controller" probeResult="failure" output=< Dec 06 16:02:10 crc kubenswrapper[4813]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 06 16:02:10 crc kubenswrapper[4813]: > Dec 06 16:02:10 crc kubenswrapper[4813]: I1206 16:02:10.386016 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-mk4nr" Dec 06 16:02:10 crc kubenswrapper[4813]: I1206 16:02:10.596433 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-bg6cb-config-gnf78"] Dec 06 16:02:10 crc kubenswrapper[4813]: E1206 16:02:10.596771 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f67c4c64-2c67-46bc-af93-6ca6ac4f4814" containerName="swift-ring-rebalance" Dec 06 16:02:10 crc kubenswrapper[4813]: I1206 16:02:10.596787 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f67c4c64-2c67-46bc-af93-6ca6ac4f4814" containerName="swift-ring-rebalance" Dec 06 16:02:10 crc kubenswrapper[4813]: I1206 16:02:10.596942 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f67c4c64-2c67-46bc-af93-6ca6ac4f4814" containerName="swift-ring-rebalance" Dec 06 16:02:10 crc kubenswrapper[4813]: I1206 16:02:10.597414 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bg6cb-config-gnf78" Dec 06 16:02:10 crc kubenswrapper[4813]: I1206 16:02:10.642453 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 06 16:02:10 crc kubenswrapper[4813]: I1206 16:02:10.651981 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-bg6cb-config-gnf78"] Dec 06 16:02:10 crc kubenswrapper[4813]: I1206 16:02:10.748950 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bglhb\" (UniqueName: \"kubernetes.io/projected/c428b428-8332-4fa6-90e5-f3a159e1d7cb-kube-api-access-bglhb\") pod \"ovn-controller-bg6cb-config-gnf78\" (UID: \"c428b428-8332-4fa6-90e5-f3a159e1d7cb\") " pod="openstack/ovn-controller-bg6cb-config-gnf78" Dec 06 16:02:10 crc kubenswrapper[4813]: I1206 16:02:10.749217 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c428b428-8332-4fa6-90e5-f3a159e1d7cb-var-run-ovn\") pod \"ovn-controller-bg6cb-config-gnf78\" (UID: \"c428b428-8332-4fa6-90e5-f3a159e1d7cb\") " pod="openstack/ovn-controller-bg6cb-config-gnf78" Dec 06 16:02:10 crc kubenswrapper[4813]: I1206 16:02:10.749286 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c428b428-8332-4fa6-90e5-f3a159e1d7cb-scripts\") pod \"ovn-controller-bg6cb-config-gnf78\" (UID: \"c428b428-8332-4fa6-90e5-f3a159e1d7cb\") " pod="openstack/ovn-controller-bg6cb-config-gnf78" Dec 06 16:02:10 crc kubenswrapper[4813]: I1206 16:02:10.749332 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c428b428-8332-4fa6-90e5-f3a159e1d7cb-var-run\") pod \"ovn-controller-bg6cb-config-gnf78\" (UID: \"c428b428-8332-4fa6-90e5-f3a159e1d7cb\") " pod="openstack/ovn-controller-bg6cb-config-gnf78" Dec 06 16:02:10 crc kubenswrapper[4813]: I1206 16:02:10.749348 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c428b428-8332-4fa6-90e5-f3a159e1d7cb-var-log-ovn\") pod \"ovn-controller-bg6cb-config-gnf78\" (UID: \"c428b428-8332-4fa6-90e5-f3a159e1d7cb\") " pod="openstack/ovn-controller-bg6cb-config-gnf78" Dec 06 16:02:10 crc kubenswrapper[4813]: I1206 16:02:10.749508 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c428b428-8332-4fa6-90e5-f3a159e1d7cb-additional-scripts\") pod \"ovn-controller-bg6cb-config-gnf78\" (UID: \"c428b428-8332-4fa6-90e5-f3a159e1d7cb\") " pod="openstack/ovn-controller-bg6cb-config-gnf78" Dec 06 16:02:10 crc kubenswrapper[4813]: I1206 16:02:10.851477 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c428b428-8332-4fa6-90e5-f3a159e1d7cb-var-run-ovn\") pod \"ovn-controller-bg6cb-config-gnf78\" (UID: \"c428b428-8332-4fa6-90e5-f3a159e1d7cb\") " pod="openstack/ovn-controller-bg6cb-config-gnf78" Dec 06 16:02:10 crc kubenswrapper[4813]: I1206 16:02:10.851539 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c428b428-8332-4fa6-90e5-f3a159e1d7cb-scripts\") pod \"ovn-controller-bg6cb-config-gnf78\" (UID: \"c428b428-8332-4fa6-90e5-f3a159e1d7cb\") " pod="openstack/ovn-controller-bg6cb-config-gnf78" Dec 06 16:02:10 crc kubenswrapper[4813]: I1206 16:02:10.851571 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c428b428-8332-4fa6-90e5-f3a159e1d7cb-var-run\") pod \"ovn-controller-bg6cb-config-gnf78\" (UID: \"c428b428-8332-4fa6-90e5-f3a159e1d7cb\") " pod="openstack/ovn-controller-bg6cb-config-gnf78" Dec 06 16:02:10 crc kubenswrapper[4813]: I1206 16:02:10.851587 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c428b428-8332-4fa6-90e5-f3a159e1d7cb-var-log-ovn\") pod \"ovn-controller-bg6cb-config-gnf78\" (UID: \"c428b428-8332-4fa6-90e5-f3a159e1d7cb\") " pod="openstack/ovn-controller-bg6cb-config-gnf78" Dec 06 16:02:10 crc kubenswrapper[4813]: I1206 16:02:10.851624 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c428b428-8332-4fa6-90e5-f3a159e1d7cb-additional-scripts\") pod \"ovn-controller-bg6cb-config-gnf78\" (UID: \"c428b428-8332-4fa6-90e5-f3a159e1d7cb\") " pod="openstack/ovn-controller-bg6cb-config-gnf78" Dec 06 16:02:10 crc kubenswrapper[4813]: I1206 16:02:10.851680 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bglhb\" (UniqueName: \"kubernetes.io/projected/c428b428-8332-4fa6-90e5-f3a159e1d7cb-kube-api-access-bglhb\") pod \"ovn-controller-bg6cb-config-gnf78\" (UID: \"c428b428-8332-4fa6-90e5-f3a159e1d7cb\") " pod="openstack/ovn-controller-bg6cb-config-gnf78" Dec 06 16:02:10 crc kubenswrapper[4813]: I1206 16:02:10.851816 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c428b428-8332-4fa6-90e5-f3a159e1d7cb-var-run-ovn\") pod \"ovn-controller-bg6cb-config-gnf78\" (UID: \"c428b428-8332-4fa6-90e5-f3a159e1d7cb\") " pod="openstack/ovn-controller-bg6cb-config-gnf78" Dec 06 16:02:10 crc kubenswrapper[4813]: I1206 16:02:10.851836 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c428b428-8332-4fa6-90e5-f3a159e1d7cb-var-run\") pod \"ovn-controller-bg6cb-config-gnf78\" (UID: \"c428b428-8332-4fa6-90e5-f3a159e1d7cb\") " pod="openstack/ovn-controller-bg6cb-config-gnf78" Dec 06 16:02:10 crc kubenswrapper[4813]: I1206 16:02:10.851829 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c428b428-8332-4fa6-90e5-f3a159e1d7cb-var-log-ovn\") pod \"ovn-controller-bg6cb-config-gnf78\" (UID: \"c428b428-8332-4fa6-90e5-f3a159e1d7cb\") " pod="openstack/ovn-controller-bg6cb-config-gnf78" Dec 06 16:02:10 crc kubenswrapper[4813]: I1206 16:02:10.852571 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c428b428-8332-4fa6-90e5-f3a159e1d7cb-additional-scripts\") pod \"ovn-controller-bg6cb-config-gnf78\" (UID: \"c428b428-8332-4fa6-90e5-f3a159e1d7cb\") " pod="openstack/ovn-controller-bg6cb-config-gnf78" Dec 06 16:02:10 crc kubenswrapper[4813]: I1206 16:02:10.854447 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c428b428-8332-4fa6-90e5-f3a159e1d7cb-scripts\") pod \"ovn-controller-bg6cb-config-gnf78\" (UID: \"c428b428-8332-4fa6-90e5-f3a159e1d7cb\") " pod="openstack/ovn-controller-bg6cb-config-gnf78" Dec 06 16:02:10 crc kubenswrapper[4813]: I1206 16:02:10.873426 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bglhb\" (UniqueName: \"kubernetes.io/projected/c428b428-8332-4fa6-90e5-f3a159e1d7cb-kube-api-access-bglhb\") pod \"ovn-controller-bg6cb-config-gnf78\" (UID: \"c428b428-8332-4fa6-90e5-f3a159e1d7cb\") " pod="openstack/ovn-controller-bg6cb-config-gnf78" Dec 06 16:02:10 crc kubenswrapper[4813]: I1206 16:02:10.962743 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bg6cb-config-gnf78" Dec 06 16:02:15 crc kubenswrapper[4813]: I1206 16:02:15.333981 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-bg6cb" podUID="ba6825c1-2137-46d6-a695-6bdd53a13ec4" containerName="ovn-controller" probeResult="failure" output=< Dec 06 16:02:15 crc kubenswrapper[4813]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 06 16:02:15 crc kubenswrapper[4813]: > Dec 06 16:02:16 crc kubenswrapper[4813]: I1206 16:02:16.529066 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-bg6cb-config-gnf78"] Dec 06 16:02:16 crc kubenswrapper[4813]: I1206 16:02:16.532147 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f571761-c8de-485f-89de-514fc0a66a55","Type":"ContainerStarted","Data":"e7fe4e2972c281257b747757733ab6aaae7926a5bf96c30d731dfc7a76fae237"} Dec 06 16:02:16 crc kubenswrapper[4813]: W1206 16:02:16.548401 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc428b428_8332_4fa6_90e5_f3a159e1d7cb.slice/crio-3c527ca2866485946a6e0589b00c84595937b1ef6525dd8dc062f983c5832399 WatchSource:0}: Error finding container 3c527ca2866485946a6e0589b00c84595937b1ef6525dd8dc062f983c5832399: Status 404 returned error can't find the container with id 3c527ca2866485946a6e0589b00c84595937b1ef6525dd8dc062f983c5832399 Dec 06 16:02:17 crc kubenswrapper[4813]: I1206 16:02:17.543567 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f571761-c8de-485f-89de-514fc0a66a55","Type":"ContainerStarted","Data":"8be505ae5807935379dc571410b484708084cc33bb1606783afb2d637a1d3c45"} Dec 06 16:02:17 crc kubenswrapper[4813]: I1206 16:02:17.545182 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tzd8d" event={"ID":"27b4a8bf-5c37-4936-8ced-6a1d02249dd1","Type":"ContainerStarted","Data":"1ef292affd710b39999c751c2f5a2ca32230e936cfa573d521c6d139e9f27c7c"} Dec 06 16:02:17 crc kubenswrapper[4813]: I1206 16:02:17.548104 4813 generic.go:334] "Generic (PLEG): container finished" podID="c428b428-8332-4fa6-90e5-f3a159e1d7cb" containerID="40897f9b7d87e956550d6ad37725608ed528e8b32b221fdab9ff83f5da018545" exitCode=0 Dec 06 16:02:17 crc kubenswrapper[4813]: I1206 16:02:17.548138 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-bg6cb-config-gnf78" event={"ID":"c428b428-8332-4fa6-90e5-f3a159e1d7cb","Type":"ContainerDied","Data":"40897f9b7d87e956550d6ad37725608ed528e8b32b221fdab9ff83f5da018545"} Dec 06 16:02:17 crc kubenswrapper[4813]: I1206 16:02:17.548545 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-bg6cb-config-gnf78" event={"ID":"c428b428-8332-4fa6-90e5-f3a159e1d7cb","Type":"ContainerStarted","Data":"3c527ca2866485946a6e0589b00c84595937b1ef6525dd8dc062f983c5832399"} Dec 06 16:02:17 crc kubenswrapper[4813]: I1206 16:02:17.564040 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-tzd8d" podStartSLOduration=3.248973943 podStartE2EDuration="18.564015334s" podCreationTimestamp="2025-12-06 16:01:59 +0000 UTC" firstStartedPulling="2025-12-06 16:02:00.779594311 +0000 UTC m=+960.670473887" lastFinishedPulling="2025-12-06 16:02:16.094635702 +0000 UTC m=+975.985515278" observedRunningTime="2025-12-06 16:02:17.561786495 +0000 UTC m=+977.452666101" watchObservedRunningTime="2025-12-06 16:02:17.564015334 +0000 UTC m=+977.454894950" Dec 06 16:02:18 crc kubenswrapper[4813]: I1206 16:02:18.563121 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f571761-c8de-485f-89de-514fc0a66a55","Type":"ContainerStarted","Data":"a68dc5c17ec7f86a26f1dfece8b7b98dcfeee1a2608cc741a501b102aa41fc09"} Dec 06 16:02:18 crc kubenswrapper[4813]: I1206 16:02:18.563374 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f571761-c8de-485f-89de-514fc0a66a55","Type":"ContainerStarted","Data":"1f9609ade4a281518f8eab72ab6f329ae25271b756faa5d053eb81bc67ad1051"} Dec 06 16:02:18 crc kubenswrapper[4813]: I1206 16:02:18.563387 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f571761-c8de-485f-89de-514fc0a66a55","Type":"ContainerStarted","Data":"0706e458f1291650808dcd3c658186e67486ac01ded59c1a30efa857748ffadd"} Dec 06 16:02:18 crc kubenswrapper[4813]: I1206 16:02:18.563397 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f571761-c8de-485f-89de-514fc0a66a55","Type":"ContainerStarted","Data":"5a5fc01821143799893bcb98ca8dc240c1980a68645d07c915d4e6b4a2ed1028"} Dec 06 16:02:18 crc kubenswrapper[4813]: I1206 16:02:18.947694 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bg6cb-config-gnf78" Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.098513 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c428b428-8332-4fa6-90e5-f3a159e1d7cb-var-run-ovn\") pod \"c428b428-8332-4fa6-90e5-f3a159e1d7cb\" (UID: \"c428b428-8332-4fa6-90e5-f3a159e1d7cb\") " Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.098575 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c428b428-8332-4fa6-90e5-f3a159e1d7cb-additional-scripts\") pod \"c428b428-8332-4fa6-90e5-f3a159e1d7cb\" (UID: \"c428b428-8332-4fa6-90e5-f3a159e1d7cb\") " Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.098624 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c428b428-8332-4fa6-90e5-f3a159e1d7cb-var-run\") pod \"c428b428-8332-4fa6-90e5-f3a159e1d7cb\" (UID: \"c428b428-8332-4fa6-90e5-f3a159e1d7cb\") " Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.098663 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c428b428-8332-4fa6-90e5-f3a159e1d7cb-scripts\") pod \"c428b428-8332-4fa6-90e5-f3a159e1d7cb\" (UID: \"c428b428-8332-4fa6-90e5-f3a159e1d7cb\") " Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.098734 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c428b428-8332-4fa6-90e5-f3a159e1d7cb-var-run" (OuterVolumeSpecName: "var-run") pod "c428b428-8332-4fa6-90e5-f3a159e1d7cb" (UID: "c428b428-8332-4fa6-90e5-f3a159e1d7cb"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.099403 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c428b428-8332-4fa6-90e5-f3a159e1d7cb-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "c428b428-8332-4fa6-90e5-f3a159e1d7cb" (UID: "c428b428-8332-4fa6-90e5-f3a159e1d7cb"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.099627 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c428b428-8332-4fa6-90e5-f3a159e1d7cb-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "c428b428-8332-4fa6-90e5-f3a159e1d7cb" (UID: "c428b428-8332-4fa6-90e5-f3a159e1d7cb"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.099670 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c428b428-8332-4fa6-90e5-f3a159e1d7cb-scripts" (OuterVolumeSpecName: "scripts") pod "c428b428-8332-4fa6-90e5-f3a159e1d7cb" (UID: "c428b428-8332-4fa6-90e5-f3a159e1d7cb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.099689 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bglhb\" (UniqueName: \"kubernetes.io/projected/c428b428-8332-4fa6-90e5-f3a159e1d7cb-kube-api-access-bglhb\") pod \"c428b428-8332-4fa6-90e5-f3a159e1d7cb\" (UID: \"c428b428-8332-4fa6-90e5-f3a159e1d7cb\") " Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.099780 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c428b428-8332-4fa6-90e5-f3a159e1d7cb-var-log-ovn\") pod \"c428b428-8332-4fa6-90e5-f3a159e1d7cb\" (UID: \"c428b428-8332-4fa6-90e5-f3a159e1d7cb\") " Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.099891 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c428b428-8332-4fa6-90e5-f3a159e1d7cb-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "c428b428-8332-4fa6-90e5-f3a159e1d7cb" (UID: "c428b428-8332-4fa6-90e5-f3a159e1d7cb"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.100074 4813 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c428b428-8332-4fa6-90e5-f3a159e1d7cb-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.100085 4813 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c428b428-8332-4fa6-90e5-f3a159e1d7cb-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.100094 4813 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c428b428-8332-4fa6-90e5-f3a159e1d7cb-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.100104 4813 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c428b428-8332-4fa6-90e5-f3a159e1d7cb-var-run\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.100114 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c428b428-8332-4fa6-90e5-f3a159e1d7cb-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.104278 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c428b428-8332-4fa6-90e5-f3a159e1d7cb-kube-api-access-bglhb" (OuterVolumeSpecName: "kube-api-access-bglhb") pod "c428b428-8332-4fa6-90e5-f3a159e1d7cb" (UID: "c428b428-8332-4fa6-90e5-f3a159e1d7cb"). InnerVolumeSpecName "kube-api-access-bglhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.201956 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bglhb\" (UniqueName: \"kubernetes.io/projected/c428b428-8332-4fa6-90e5-f3a159e1d7cb-kube-api-access-bglhb\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.427607 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.427666 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.573538 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-bg6cb-config-gnf78" event={"ID":"c428b428-8332-4fa6-90e5-f3a159e1d7cb","Type":"ContainerDied","Data":"3c527ca2866485946a6e0589b00c84595937b1ef6525dd8dc062f983c5832399"} Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.573594 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c527ca2866485946a6e0589b00c84595937b1ef6525dd8dc062f983c5832399" Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.574780 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bg6cb-config-gnf78" Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.580965 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f571761-c8de-485f-89de-514fc0a66a55","Type":"ContainerStarted","Data":"eab7df3ed5d87c8b9461988af83bb3b07d0e9a6bb408dcf1deab5039ffc5f9b6"} Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.581006 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f571761-c8de-485f-89de-514fc0a66a55","Type":"ContainerStarted","Data":"627e7f52e4b74687cebb6de87fc018e986d7875d08d1324fa20abc55bea0c5ba"} Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.581016 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f571761-c8de-485f-89de-514fc0a66a55","Type":"ContainerStarted","Data":"187555f662c89e26d8dd1472a43e643cd2393e86838631eeeb55ea5976cb05a5"} Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.644930 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=20.526594299 podStartE2EDuration="34.644913723s" podCreationTimestamp="2025-12-06 16:01:45 +0000 UTC" firstStartedPulling="2025-12-06 16:02:03.56810929 +0000 UTC m=+963.458988866" lastFinishedPulling="2025-12-06 16:02:17.686428704 +0000 UTC m=+977.577308290" observedRunningTime="2025-12-06 16:02:19.638556365 +0000 UTC m=+979.529435961" watchObservedRunningTime="2025-12-06 16:02:19.644913723 +0000 UTC m=+979.535793309" Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.975607 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-tbd65"] Dec 06 16:02:19 crc kubenswrapper[4813]: E1206 16:02:19.975921 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c428b428-8332-4fa6-90e5-f3a159e1d7cb" containerName="ovn-config" Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.975937 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="c428b428-8332-4fa6-90e5-f3a159e1d7cb" containerName="ovn-config" Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.976101 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="c428b428-8332-4fa6-90e5-f3a159e1d7cb" containerName="ovn-config" Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.976923 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" Dec 06 16:02:19 crc kubenswrapper[4813]: I1206 16:02:19.978768 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 06 16:02:20 crc kubenswrapper[4813]: I1206 16:02:20.003173 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-tbd65"] Dec 06 16:02:20 crc kubenswrapper[4813]: I1206 16:02:20.108594 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-bg6cb-config-gnf78"] Dec 06 16:02:20 crc kubenswrapper[4813]: I1206 16:02:20.119317 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-bg6cb-config-gnf78"] Dec 06 16:02:20 crc kubenswrapper[4813]: I1206 16:02:20.131371 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/274211c2-377d-4164-be60-ca42e8167849-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-tbd65\" (UID: \"274211c2-377d-4164-be60-ca42e8167849\") " pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" Dec 06 16:02:20 crc kubenswrapper[4813]: I1206 16:02:20.131447 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/274211c2-377d-4164-be60-ca42e8167849-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-tbd65\" (UID: \"274211c2-377d-4164-be60-ca42e8167849\") " pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" Dec 06 16:02:20 crc kubenswrapper[4813]: I1206 16:02:20.131574 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/274211c2-377d-4164-be60-ca42e8167849-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-tbd65\" (UID: \"274211c2-377d-4164-be60-ca42e8167849\") " pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" Dec 06 16:02:20 crc kubenswrapper[4813]: I1206 16:02:20.131669 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lclqs\" (UniqueName: \"kubernetes.io/projected/274211c2-377d-4164-be60-ca42e8167849-kube-api-access-lclqs\") pod \"dnsmasq-dns-5c79d794d7-tbd65\" (UID: \"274211c2-377d-4164-be60-ca42e8167849\") " pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" Dec 06 16:02:20 crc kubenswrapper[4813]: I1206 16:02:20.131709 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/274211c2-377d-4164-be60-ca42e8167849-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-tbd65\" (UID: \"274211c2-377d-4164-be60-ca42e8167849\") " pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" Dec 06 16:02:20 crc kubenswrapper[4813]: I1206 16:02:20.131755 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/274211c2-377d-4164-be60-ca42e8167849-config\") pod \"dnsmasq-dns-5c79d794d7-tbd65\" (UID: \"274211c2-377d-4164-be60-ca42e8167849\") " pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" Dec 06 16:02:20 crc kubenswrapper[4813]: I1206 16:02:20.233958 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/274211c2-377d-4164-be60-ca42e8167849-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-tbd65\" (UID: \"274211c2-377d-4164-be60-ca42e8167849\") " pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" Dec 06 16:02:20 crc kubenswrapper[4813]: I1206 16:02:20.234005 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lclqs\" (UniqueName: \"kubernetes.io/projected/274211c2-377d-4164-be60-ca42e8167849-kube-api-access-lclqs\") pod \"dnsmasq-dns-5c79d794d7-tbd65\" (UID: \"274211c2-377d-4164-be60-ca42e8167849\") " pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" Dec 06 16:02:20 crc kubenswrapper[4813]: I1206 16:02:20.234041 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/274211c2-377d-4164-be60-ca42e8167849-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-tbd65\" (UID: \"274211c2-377d-4164-be60-ca42e8167849\") " pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" Dec 06 16:02:20 crc kubenswrapper[4813]: I1206 16:02:20.234085 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/274211c2-377d-4164-be60-ca42e8167849-config\") pod \"dnsmasq-dns-5c79d794d7-tbd65\" (UID: \"274211c2-377d-4164-be60-ca42e8167849\") " pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" Dec 06 16:02:20 crc kubenswrapper[4813]: I1206 16:02:20.234140 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/274211c2-377d-4164-be60-ca42e8167849-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-tbd65\" (UID: \"274211c2-377d-4164-be60-ca42e8167849\") " pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" Dec 06 16:02:20 crc kubenswrapper[4813]: I1206 16:02:20.234164 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/274211c2-377d-4164-be60-ca42e8167849-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-tbd65\" (UID: \"274211c2-377d-4164-be60-ca42e8167849\") " pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" Dec 06 16:02:20 crc kubenswrapper[4813]: I1206 16:02:20.234992 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/274211c2-377d-4164-be60-ca42e8167849-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-tbd65\" (UID: \"274211c2-377d-4164-be60-ca42e8167849\") " pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" Dec 06 16:02:20 crc kubenswrapper[4813]: I1206 16:02:20.235509 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/274211c2-377d-4164-be60-ca42e8167849-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-tbd65\" (UID: \"274211c2-377d-4164-be60-ca42e8167849\") " pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" Dec 06 16:02:20 crc kubenswrapper[4813]: I1206 16:02:20.236274 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/274211c2-377d-4164-be60-ca42e8167849-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-tbd65\" (UID: \"274211c2-377d-4164-be60-ca42e8167849\") " pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" Dec 06 16:02:20 crc kubenswrapper[4813]: I1206 16:02:20.236761 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/274211c2-377d-4164-be60-ca42e8167849-config\") pod \"dnsmasq-dns-5c79d794d7-tbd65\" (UID: \"274211c2-377d-4164-be60-ca42e8167849\") " pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" Dec 06 16:02:20 crc kubenswrapper[4813]: I1206 16:02:20.237330 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/274211c2-377d-4164-be60-ca42e8167849-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-tbd65\" (UID: \"274211c2-377d-4164-be60-ca42e8167849\") " pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" Dec 06 16:02:20 crc kubenswrapper[4813]: I1206 16:02:20.254596 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lclqs\" (UniqueName: \"kubernetes.io/projected/274211c2-377d-4164-be60-ca42e8167849-kube-api-access-lclqs\") pod \"dnsmasq-dns-5c79d794d7-tbd65\" (UID: \"274211c2-377d-4164-be60-ca42e8167849\") " pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" Dec 06 16:02:20 crc kubenswrapper[4813]: I1206 16:02:20.290152 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" Dec 06 16:02:20 crc kubenswrapper[4813]: I1206 16:02:20.339464 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:02:20 crc kubenswrapper[4813]: I1206 16:02:20.351544 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-bg6cb" Dec 06 16:02:20 crc kubenswrapper[4813]: I1206 16:02:20.510675 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c428b428-8332-4fa6-90e5-f3a159e1d7cb" path="/var/lib/kubelet/pods/c428b428-8332-4fa6-90e5-f3a159e1d7cb/volumes" Dec 06 16:02:20 crc kubenswrapper[4813]: I1206 16:02:20.689926 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 06 16:02:20 crc kubenswrapper[4813]: I1206 16:02:20.852916 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-tbd65"] Dec 06 16:02:21 crc kubenswrapper[4813]: I1206 16:02:21.602327 4813 generic.go:334] "Generic (PLEG): container finished" podID="274211c2-377d-4164-be60-ca42e8167849" containerID="3f280ff8667572d95419dbd6b782b0534fca40eb5d29d4cd3b6d0457f3a49c36" exitCode=0 Dec 06 16:02:21 crc kubenswrapper[4813]: I1206 16:02:21.602471 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" event={"ID":"274211c2-377d-4164-be60-ca42e8167849","Type":"ContainerDied","Data":"3f280ff8667572d95419dbd6b782b0534fca40eb5d29d4cd3b6d0457f3a49c36"} Dec 06 16:02:21 crc kubenswrapper[4813]: I1206 16:02:21.602941 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" event={"ID":"274211c2-377d-4164-be60-ca42e8167849","Type":"ContainerStarted","Data":"72cbdf2a26fc9e3245239a0b45e78308483faa2232e2b22236f03bc54e32b645"} Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.319325 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-88qpf"] Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.321576 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-88qpf" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.368818 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60776264-205f-4541-8d51-3d0d3aec891d-operator-scripts\") pod \"cinder-db-create-88qpf\" (UID: \"60776264-205f-4541-8d51-3d0d3aec891d\") " pod="openstack/cinder-db-create-88qpf" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.368939 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwpk7\" (UniqueName: \"kubernetes.io/projected/60776264-205f-4541-8d51-3d0d3aec891d-kube-api-access-gwpk7\") pod \"cinder-db-create-88qpf\" (UID: \"60776264-205f-4541-8d51-3d0d3aec891d\") " pod="openstack/cinder-db-create-88qpf" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.425539 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-88qpf"] Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.436534 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-4flrx"] Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.437623 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4flrx" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.443534 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-2f15-account-create-update-2gvw4"] Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.444523 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-2f15-account-create-update-2gvw4" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.445900 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.470817 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwpk7\" (UniqueName: \"kubernetes.io/projected/60776264-205f-4541-8d51-3d0d3aec891d-kube-api-access-gwpk7\") pod \"cinder-db-create-88qpf\" (UID: \"60776264-205f-4541-8d51-3d0d3aec891d\") " pod="openstack/cinder-db-create-88qpf" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.471358 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cggz6\" (UniqueName: \"kubernetes.io/projected/235e0c8b-b891-4c0c-9676-d0df7a2b6836-kube-api-access-cggz6\") pod \"barbican-db-create-4flrx\" (UID: \"235e0c8b-b891-4c0c-9676-d0df7a2b6836\") " pod="openstack/barbican-db-create-4flrx" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.471473 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60776264-205f-4541-8d51-3d0d3aec891d-operator-scripts\") pod \"cinder-db-create-88qpf\" (UID: \"60776264-205f-4541-8d51-3d0d3aec891d\") " pod="openstack/cinder-db-create-88qpf" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.472213 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/235e0c8b-b891-4c0c-9676-d0df7a2b6836-operator-scripts\") pod \"barbican-db-create-4flrx\" (UID: \"235e0c8b-b891-4c0c-9676-d0df7a2b6836\") " pod="openstack/barbican-db-create-4flrx" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.472169 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60776264-205f-4541-8d51-3d0d3aec891d-operator-scripts\") pod \"cinder-db-create-88qpf\" (UID: \"60776264-205f-4541-8d51-3d0d3aec891d\") " pod="openstack/cinder-db-create-88qpf" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.499416 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-4flrx"] Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.523094 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwpk7\" (UniqueName: \"kubernetes.io/projected/60776264-205f-4541-8d51-3d0d3aec891d-kube-api-access-gwpk7\") pod \"cinder-db-create-88qpf\" (UID: \"60776264-205f-4541-8d51-3d0d3aec891d\") " pod="openstack/cinder-db-create-88qpf" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.544305 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-2f15-account-create-update-2gvw4"] Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.573935 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cggz6\" (UniqueName: \"kubernetes.io/projected/235e0c8b-b891-4c0c-9676-d0df7a2b6836-kube-api-access-cggz6\") pod \"barbican-db-create-4flrx\" (UID: \"235e0c8b-b891-4c0c-9676-d0df7a2b6836\") " pod="openstack/barbican-db-create-4flrx" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.573980 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vn6jh\" (UniqueName: \"kubernetes.io/projected/612d9b68-b55f-40f5-a45a-c1d7a3b32f2b-kube-api-access-vn6jh\") pod \"cinder-2f15-account-create-update-2gvw4\" (UID: \"612d9b68-b55f-40f5-a45a-c1d7a3b32f2b\") " pod="openstack/cinder-2f15-account-create-update-2gvw4" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.574027 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/235e0c8b-b891-4c0c-9676-d0df7a2b6836-operator-scripts\") pod \"barbican-db-create-4flrx\" (UID: \"235e0c8b-b891-4c0c-9676-d0df7a2b6836\") " pod="openstack/barbican-db-create-4flrx" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.574085 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/612d9b68-b55f-40f5-a45a-c1d7a3b32f2b-operator-scripts\") pod \"cinder-2f15-account-create-update-2gvw4\" (UID: \"612d9b68-b55f-40f5-a45a-c1d7a3b32f2b\") " pod="openstack/cinder-2f15-account-create-update-2gvw4" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.574716 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/235e0c8b-b891-4c0c-9676-d0df7a2b6836-operator-scripts\") pod \"barbican-db-create-4flrx\" (UID: \"235e0c8b-b891-4c0c-9676-d0df7a2b6836\") " pod="openstack/barbican-db-create-4flrx" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.604648 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cggz6\" (UniqueName: \"kubernetes.io/projected/235e0c8b-b891-4c0c-9676-d0df7a2b6836-kube-api-access-cggz6\") pod \"barbican-db-create-4flrx\" (UID: \"235e0c8b-b891-4c0c-9676-d0df7a2b6836\") " pod="openstack/barbican-db-create-4flrx" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.614978 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" event={"ID":"274211c2-377d-4164-be60-ca42e8167849","Type":"ContainerStarted","Data":"554a49704af221c2f375a789f3ae52a335b47c491aacfa5eacdbc02f838c3401"} Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.615988 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.627068 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-24e9-account-create-update-ht2lg"] Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.628108 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-24e9-account-create-update-ht2lg" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.648905 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-88qpf" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.651411 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.675576 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/612d9b68-b55f-40f5-a45a-c1d7a3b32f2b-operator-scripts\") pod \"cinder-2f15-account-create-update-2gvw4\" (UID: \"612d9b68-b55f-40f5-a45a-c1d7a3b32f2b\") " pod="openstack/cinder-2f15-account-create-update-2gvw4" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.675798 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9-operator-scripts\") pod \"barbican-24e9-account-create-update-ht2lg\" (UID: \"9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9\") " pod="openstack/barbican-24e9-account-create-update-ht2lg" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.675927 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbp8s\" (UniqueName: \"kubernetes.io/projected/9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9-kube-api-access-nbp8s\") pod \"barbican-24e9-account-create-update-ht2lg\" (UID: \"9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9\") " pod="openstack/barbican-24e9-account-create-update-ht2lg" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.676021 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vn6jh\" (UniqueName: \"kubernetes.io/projected/612d9b68-b55f-40f5-a45a-c1d7a3b32f2b-kube-api-access-vn6jh\") pod \"cinder-2f15-account-create-update-2gvw4\" (UID: \"612d9b68-b55f-40f5-a45a-c1d7a3b32f2b\") " pod="openstack/cinder-2f15-account-create-update-2gvw4" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.676252 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/612d9b68-b55f-40f5-a45a-c1d7a3b32f2b-operator-scripts\") pod \"cinder-2f15-account-create-update-2gvw4\" (UID: \"612d9b68-b55f-40f5-a45a-c1d7a3b32f2b\") " pod="openstack/cinder-2f15-account-create-update-2gvw4" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.735844 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-24e9-account-create-update-ht2lg"] Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.753679 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4flrx" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.770803 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" podStartSLOduration=3.770784311 podStartE2EDuration="3.770784311s" podCreationTimestamp="2025-12-06 16:02:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:02:22.735532328 +0000 UTC m=+982.626411904" watchObservedRunningTime="2025-12-06 16:02:22.770784311 +0000 UTC m=+982.661663887" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.781362 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9-operator-scripts\") pod \"barbican-24e9-account-create-update-ht2lg\" (UID: \"9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9\") " pod="openstack/barbican-24e9-account-create-update-ht2lg" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.781432 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbp8s\" (UniqueName: \"kubernetes.io/projected/9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9-kube-api-access-nbp8s\") pod \"barbican-24e9-account-create-update-ht2lg\" (UID: \"9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9\") " pod="openstack/barbican-24e9-account-create-update-ht2lg" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.782368 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9-operator-scripts\") pod \"barbican-24e9-account-create-update-ht2lg\" (UID: \"9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9\") " pod="openstack/barbican-24e9-account-create-update-ht2lg" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.818916 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vn6jh\" (UniqueName: \"kubernetes.io/projected/612d9b68-b55f-40f5-a45a-c1d7a3b32f2b-kube-api-access-vn6jh\") pod \"cinder-2f15-account-create-update-2gvw4\" (UID: \"612d9b68-b55f-40f5-a45a-c1d7a3b32f2b\") " pod="openstack/cinder-2f15-account-create-update-2gvw4" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.867578 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbp8s\" (UniqueName: \"kubernetes.io/projected/9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9-kube-api-access-nbp8s\") pod \"barbican-24e9-account-create-update-ht2lg\" (UID: \"9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9\") " pod="openstack/barbican-24e9-account-create-update-ht2lg" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.870327 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-6nm8x"] Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.872181 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-6nm8x" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.898866 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-6nm8x"] Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.955321 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-c831-account-create-update-h8lzr"] Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.956423 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c831-account-create-update-h8lzr" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.960450 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.968408 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-24e9-account-create-update-ht2lg" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.974600 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c831-account-create-update-h8lzr"] Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.994147 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khnfv\" (UniqueName: \"kubernetes.io/projected/3579ce45-95fa-44af-a164-3153a1c873ac-kube-api-access-khnfv\") pod \"neutron-c831-account-create-update-h8lzr\" (UID: \"3579ce45-95fa-44af-a164-3153a1c873ac\") " pod="openstack/neutron-c831-account-create-update-h8lzr" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.994212 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3579ce45-95fa-44af-a164-3153a1c873ac-operator-scripts\") pod \"neutron-c831-account-create-update-h8lzr\" (UID: \"3579ce45-95fa-44af-a164-3153a1c873ac\") " pod="openstack/neutron-c831-account-create-update-h8lzr" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.994233 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/93edeafb-90c8-4ede-b6f4-608b67aadc37-operator-scripts\") pod \"neutron-db-create-6nm8x\" (UID: \"93edeafb-90c8-4ede-b6f4-608b67aadc37\") " pod="openstack/neutron-db-create-6nm8x" Dec 06 16:02:22 crc kubenswrapper[4813]: I1206 16:02:22.994278 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxjbb\" (UniqueName: \"kubernetes.io/projected/93edeafb-90c8-4ede-b6f4-608b67aadc37-kube-api-access-bxjbb\") pod \"neutron-db-create-6nm8x\" (UID: \"93edeafb-90c8-4ede-b6f4-608b67aadc37\") " pod="openstack/neutron-db-create-6nm8x" Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.029356 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-jz5zm"] Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.032466 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jz5zm" Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.037705 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.037855 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fmvk2" Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.037986 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.038079 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.051644 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-jz5zm"] Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.062927 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-2f15-account-create-update-2gvw4" Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.100934 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cc7028f-2019-4d85-84d9-7e7bb05216ec-combined-ca-bundle\") pod \"keystone-db-sync-jz5zm\" (UID: \"4cc7028f-2019-4d85-84d9-7e7bb05216ec\") " pod="openstack/keystone-db-sync-jz5zm" Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.101057 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khnfv\" (UniqueName: \"kubernetes.io/projected/3579ce45-95fa-44af-a164-3153a1c873ac-kube-api-access-khnfv\") pod \"neutron-c831-account-create-update-h8lzr\" (UID: \"3579ce45-95fa-44af-a164-3153a1c873ac\") " pod="openstack/neutron-c831-account-create-update-h8lzr" Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.101092 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4cc7028f-2019-4d85-84d9-7e7bb05216ec-config-data\") pod \"keystone-db-sync-jz5zm\" (UID: \"4cc7028f-2019-4d85-84d9-7e7bb05216ec\") " pod="openstack/keystone-db-sync-jz5zm" Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.101137 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3579ce45-95fa-44af-a164-3153a1c873ac-operator-scripts\") pod \"neutron-c831-account-create-update-h8lzr\" (UID: \"3579ce45-95fa-44af-a164-3153a1c873ac\") " pod="openstack/neutron-c831-account-create-update-h8lzr" Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.101153 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/93edeafb-90c8-4ede-b6f4-608b67aadc37-operator-scripts\") pod \"neutron-db-create-6nm8x\" (UID: \"93edeafb-90c8-4ede-b6f4-608b67aadc37\") " pod="openstack/neutron-db-create-6nm8x" Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.101186 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxjbb\" (UniqueName: \"kubernetes.io/projected/93edeafb-90c8-4ede-b6f4-608b67aadc37-kube-api-access-bxjbb\") pod \"neutron-db-create-6nm8x\" (UID: \"93edeafb-90c8-4ede-b6f4-608b67aadc37\") " pod="openstack/neutron-db-create-6nm8x" Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.101219 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rq4zm\" (UniqueName: \"kubernetes.io/projected/4cc7028f-2019-4d85-84d9-7e7bb05216ec-kube-api-access-rq4zm\") pod \"keystone-db-sync-jz5zm\" (UID: \"4cc7028f-2019-4d85-84d9-7e7bb05216ec\") " pod="openstack/keystone-db-sync-jz5zm" Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.102065 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/93edeafb-90c8-4ede-b6f4-608b67aadc37-operator-scripts\") pod \"neutron-db-create-6nm8x\" (UID: \"93edeafb-90c8-4ede-b6f4-608b67aadc37\") " pod="openstack/neutron-db-create-6nm8x" Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.102159 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3579ce45-95fa-44af-a164-3153a1c873ac-operator-scripts\") pod \"neutron-c831-account-create-update-h8lzr\" (UID: \"3579ce45-95fa-44af-a164-3153a1c873ac\") " pod="openstack/neutron-c831-account-create-update-h8lzr" Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.130318 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxjbb\" (UniqueName: \"kubernetes.io/projected/93edeafb-90c8-4ede-b6f4-608b67aadc37-kube-api-access-bxjbb\") pod \"neutron-db-create-6nm8x\" (UID: \"93edeafb-90c8-4ede-b6f4-608b67aadc37\") " pod="openstack/neutron-db-create-6nm8x" Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.163877 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khnfv\" (UniqueName: \"kubernetes.io/projected/3579ce45-95fa-44af-a164-3153a1c873ac-kube-api-access-khnfv\") pod \"neutron-c831-account-create-update-h8lzr\" (UID: \"3579ce45-95fa-44af-a164-3153a1c873ac\") " pod="openstack/neutron-c831-account-create-update-h8lzr" Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.202195 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4cc7028f-2019-4d85-84d9-7e7bb05216ec-config-data\") pod \"keystone-db-sync-jz5zm\" (UID: \"4cc7028f-2019-4d85-84d9-7e7bb05216ec\") " pod="openstack/keystone-db-sync-jz5zm" Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.202286 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rq4zm\" (UniqueName: \"kubernetes.io/projected/4cc7028f-2019-4d85-84d9-7e7bb05216ec-kube-api-access-rq4zm\") pod \"keystone-db-sync-jz5zm\" (UID: \"4cc7028f-2019-4d85-84d9-7e7bb05216ec\") " pod="openstack/keystone-db-sync-jz5zm" Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.202322 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cc7028f-2019-4d85-84d9-7e7bb05216ec-combined-ca-bundle\") pod \"keystone-db-sync-jz5zm\" (UID: \"4cc7028f-2019-4d85-84d9-7e7bb05216ec\") " pod="openstack/keystone-db-sync-jz5zm" Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.208278 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4cc7028f-2019-4d85-84d9-7e7bb05216ec-config-data\") pod \"keystone-db-sync-jz5zm\" (UID: \"4cc7028f-2019-4d85-84d9-7e7bb05216ec\") " pod="openstack/keystone-db-sync-jz5zm" Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.208787 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cc7028f-2019-4d85-84d9-7e7bb05216ec-combined-ca-bundle\") pod \"keystone-db-sync-jz5zm\" (UID: \"4cc7028f-2019-4d85-84d9-7e7bb05216ec\") " pod="openstack/keystone-db-sync-jz5zm" Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.210618 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-6nm8x" Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.227691 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rq4zm\" (UniqueName: \"kubernetes.io/projected/4cc7028f-2019-4d85-84d9-7e7bb05216ec-kube-api-access-rq4zm\") pod \"keystone-db-sync-jz5zm\" (UID: \"4cc7028f-2019-4d85-84d9-7e7bb05216ec\") " pod="openstack/keystone-db-sync-jz5zm" Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.297635 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c831-account-create-update-h8lzr" Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.376741 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jz5zm" Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.424793 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-4flrx"] Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.623538 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4flrx" event={"ID":"235e0c8b-b891-4c0c-9676-d0df7a2b6836","Type":"ContainerStarted","Data":"f8f895887285a01de2487b83a3824161b7731aba899f19be687c15b051cb155c"} Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.713768 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-88qpf"] Dec 06 16:02:23 crc kubenswrapper[4813]: W1206 16:02:23.724820 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60776264_205f_4541_8d51_3d0d3aec891d.slice/crio-a4489a86a2e5c777635390646f071883caa0088ddd0c0d98c90d7b12d1072339 WatchSource:0}: Error finding container a4489a86a2e5c777635390646f071883caa0088ddd0c0d98c90d7b12d1072339: Status 404 returned error can't find the container with id a4489a86a2e5c777635390646f071883caa0088ddd0c0d98c90d7b12d1072339 Dec 06 16:02:23 crc kubenswrapper[4813]: I1206 16:02:23.927840 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-6nm8x"] Dec 06 16:02:24 crc kubenswrapper[4813]: I1206 16:02:24.003193 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-24e9-account-create-update-ht2lg"] Dec 06 16:02:24 crc kubenswrapper[4813]: I1206 16:02:24.052175 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-2f15-account-create-update-2gvw4"] Dec 06 16:02:24 crc kubenswrapper[4813]: W1206 16:02:24.068753 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod612d9b68_b55f_40f5_a45a_c1d7a3b32f2b.slice/crio-cffcc9065a4a92e0bce0108f3aff0758ec64a8aa52e73efdadbe9b5cd16dcede WatchSource:0}: Error finding container cffcc9065a4a92e0bce0108f3aff0758ec64a8aa52e73efdadbe9b5cd16dcede: Status 404 returned error can't find the container with id cffcc9065a4a92e0bce0108f3aff0758ec64a8aa52e73efdadbe9b5cd16dcede Dec 06 16:02:24 crc kubenswrapper[4813]: I1206 16:02:24.219579 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c831-account-create-update-h8lzr"] Dec 06 16:02:24 crc kubenswrapper[4813]: W1206 16:02:24.237884 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3579ce45_95fa_44af_a164_3153a1c873ac.slice/crio-9902d3fcf6cfcaf401d10999ba5f1dcc1b49e565991c54b9b050bc8dce029ded WatchSource:0}: Error finding container 9902d3fcf6cfcaf401d10999ba5f1dcc1b49e565991c54b9b050bc8dce029ded: Status 404 returned error can't find the container with id 9902d3fcf6cfcaf401d10999ba5f1dcc1b49e565991c54b9b050bc8dce029ded Dec 06 16:02:24 crc kubenswrapper[4813]: I1206 16:02:24.241103 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-jz5zm"] Dec 06 16:02:24 crc kubenswrapper[4813]: W1206 16:02:24.271321 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4cc7028f_2019_4d85_84d9_7e7bb05216ec.slice/crio-86e889e88c2aa449fc20ce62cef21ab9dcae0677a68ce8fcca2b63445d618a1d WatchSource:0}: Error finding container 86e889e88c2aa449fc20ce62cef21ab9dcae0677a68ce8fcca2b63445d618a1d: Status 404 returned error can't find the container with id 86e889e88c2aa449fc20ce62cef21ab9dcae0677a68ce8fcca2b63445d618a1d Dec 06 16:02:24 crc kubenswrapper[4813]: I1206 16:02:24.634727 4813 generic.go:334] "Generic (PLEG): container finished" podID="60776264-205f-4541-8d51-3d0d3aec891d" containerID="9b4593a84e5a1d8e2f4d1238ec7d3cb54471974e3cf6c166f3e9292376acdf9c" exitCode=0 Dec 06 16:02:24 crc kubenswrapper[4813]: I1206 16:02:24.634793 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-88qpf" event={"ID":"60776264-205f-4541-8d51-3d0d3aec891d","Type":"ContainerDied","Data":"9b4593a84e5a1d8e2f4d1238ec7d3cb54471974e3cf6c166f3e9292376acdf9c"} Dec 06 16:02:24 crc kubenswrapper[4813]: I1206 16:02:24.634820 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-88qpf" event={"ID":"60776264-205f-4541-8d51-3d0d3aec891d","Type":"ContainerStarted","Data":"a4489a86a2e5c777635390646f071883caa0088ddd0c0d98c90d7b12d1072339"} Dec 06 16:02:24 crc kubenswrapper[4813]: I1206 16:02:24.637957 4813 generic.go:334] "Generic (PLEG): container finished" podID="93edeafb-90c8-4ede-b6f4-608b67aadc37" containerID="7cb1f4efb714e006d797c6e6963e82be21e1b20f8c0a960ee6d4c452adb052c2" exitCode=0 Dec 06 16:02:24 crc kubenswrapper[4813]: I1206 16:02:24.638002 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-6nm8x" event={"ID":"93edeafb-90c8-4ede-b6f4-608b67aadc37","Type":"ContainerDied","Data":"7cb1f4efb714e006d797c6e6963e82be21e1b20f8c0a960ee6d4c452adb052c2"} Dec 06 16:02:24 crc kubenswrapper[4813]: I1206 16:02:24.638019 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-6nm8x" event={"ID":"93edeafb-90c8-4ede-b6f4-608b67aadc37","Type":"ContainerStarted","Data":"5833b8ecfedb13fb375d204948cf8b060a974af2ecf6223fcbbefe0dd9679c42"} Dec 06 16:02:24 crc kubenswrapper[4813]: I1206 16:02:24.645468 4813 generic.go:334] "Generic (PLEG): container finished" podID="235e0c8b-b891-4c0c-9676-d0df7a2b6836" containerID="efb2db4848d9528f745baf4875050797406d985b15dee0813fe2f32e0ee85ec3" exitCode=0 Dec 06 16:02:24 crc kubenswrapper[4813]: I1206 16:02:24.645533 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4flrx" event={"ID":"235e0c8b-b891-4c0c-9676-d0df7a2b6836","Type":"ContainerDied","Data":"efb2db4848d9528f745baf4875050797406d985b15dee0813fe2f32e0ee85ec3"} Dec 06 16:02:24 crc kubenswrapper[4813]: I1206 16:02:24.649568 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c831-account-create-update-h8lzr" event={"ID":"3579ce45-95fa-44af-a164-3153a1c873ac","Type":"ContainerStarted","Data":"154a8dd12aee5a84bc8121d41fa15b0771b23a193b0423a6ed68c69d2503a638"} Dec 06 16:02:24 crc kubenswrapper[4813]: I1206 16:02:24.649602 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c831-account-create-update-h8lzr" event={"ID":"3579ce45-95fa-44af-a164-3153a1c873ac","Type":"ContainerStarted","Data":"9902d3fcf6cfcaf401d10999ba5f1dcc1b49e565991c54b9b050bc8dce029ded"} Dec 06 16:02:24 crc kubenswrapper[4813]: I1206 16:02:24.654647 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-24e9-account-create-update-ht2lg" event={"ID":"9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9","Type":"ContainerStarted","Data":"173f5192c5dbe3f5cea4ce51a3efa20d5cdf784654ac5b83dc8ccaeaa6f592cb"} Dec 06 16:02:24 crc kubenswrapper[4813]: I1206 16:02:24.654690 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-24e9-account-create-update-ht2lg" event={"ID":"9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9","Type":"ContainerStarted","Data":"6a968d37f1afc4a031f2ca5c62b38b29d9e027398e9f66f0b3fef38ca3257560"} Dec 06 16:02:24 crc kubenswrapper[4813]: I1206 16:02:24.664882 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-2f15-account-create-update-2gvw4" event={"ID":"612d9b68-b55f-40f5-a45a-c1d7a3b32f2b","Type":"ContainerStarted","Data":"31c147352c0d0f036388b6a93a22eceb442d0473ce52fc96e9d6e849d15279c2"} Dec 06 16:02:24 crc kubenswrapper[4813]: I1206 16:02:24.664932 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-2f15-account-create-update-2gvw4" event={"ID":"612d9b68-b55f-40f5-a45a-c1d7a3b32f2b","Type":"ContainerStarted","Data":"cffcc9065a4a92e0bce0108f3aff0758ec64a8aa52e73efdadbe9b5cd16dcede"} Dec 06 16:02:24 crc kubenswrapper[4813]: I1206 16:02:24.667253 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jz5zm" event={"ID":"4cc7028f-2019-4d85-84d9-7e7bb05216ec","Type":"ContainerStarted","Data":"86e889e88c2aa449fc20ce62cef21ab9dcae0677a68ce8fcca2b63445d618a1d"} Dec 06 16:02:24 crc kubenswrapper[4813]: I1206 16:02:24.713625 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-c831-account-create-update-h8lzr" podStartSLOduration=2.713607025 podStartE2EDuration="2.713607025s" podCreationTimestamp="2025-12-06 16:02:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:02:24.687447553 +0000 UTC m=+984.578327129" watchObservedRunningTime="2025-12-06 16:02:24.713607025 +0000 UTC m=+984.604486601" Dec 06 16:02:24 crc kubenswrapper[4813]: I1206 16:02:24.718532 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-24e9-account-create-update-ht2lg" podStartSLOduration=2.718519105 podStartE2EDuration="2.718519105s" podCreationTimestamp="2025-12-06 16:02:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:02:24.711224302 +0000 UTC m=+984.602103878" watchObservedRunningTime="2025-12-06 16:02:24.718519105 +0000 UTC m=+984.609398681" Dec 06 16:02:24 crc kubenswrapper[4813]: I1206 16:02:24.791504 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-2f15-account-create-update-2gvw4" podStartSLOduration=2.791482456 podStartE2EDuration="2.791482456s" podCreationTimestamp="2025-12-06 16:02:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:02:24.756557202 +0000 UTC m=+984.647436778" watchObservedRunningTime="2025-12-06 16:02:24.791482456 +0000 UTC m=+984.682362032" Dec 06 16:02:25 crc kubenswrapper[4813]: I1206 16:02:25.676393 4813 generic.go:334] "Generic (PLEG): container finished" podID="3579ce45-95fa-44af-a164-3153a1c873ac" containerID="154a8dd12aee5a84bc8121d41fa15b0771b23a193b0423a6ed68c69d2503a638" exitCode=0 Dec 06 16:02:25 crc kubenswrapper[4813]: I1206 16:02:25.676750 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c831-account-create-update-h8lzr" event={"ID":"3579ce45-95fa-44af-a164-3153a1c873ac","Type":"ContainerDied","Data":"154a8dd12aee5a84bc8121d41fa15b0771b23a193b0423a6ed68c69d2503a638"} Dec 06 16:02:25 crc kubenswrapper[4813]: I1206 16:02:25.680777 4813 generic.go:334] "Generic (PLEG): container finished" podID="9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9" containerID="173f5192c5dbe3f5cea4ce51a3efa20d5cdf784654ac5b83dc8ccaeaa6f592cb" exitCode=0 Dec 06 16:02:25 crc kubenswrapper[4813]: I1206 16:02:25.680877 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-24e9-account-create-update-ht2lg" event={"ID":"9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9","Type":"ContainerDied","Data":"173f5192c5dbe3f5cea4ce51a3efa20d5cdf784654ac5b83dc8ccaeaa6f592cb"} Dec 06 16:02:25 crc kubenswrapper[4813]: I1206 16:02:25.683679 4813 generic.go:334] "Generic (PLEG): container finished" podID="612d9b68-b55f-40f5-a45a-c1d7a3b32f2b" containerID="31c147352c0d0f036388b6a93a22eceb442d0473ce52fc96e9d6e849d15279c2" exitCode=0 Dec 06 16:02:25 crc kubenswrapper[4813]: I1206 16:02:25.683838 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-2f15-account-create-update-2gvw4" event={"ID":"612d9b68-b55f-40f5-a45a-c1d7a3b32f2b","Type":"ContainerDied","Data":"31c147352c0d0f036388b6a93a22eceb442d0473ce52fc96e9d6e849d15279c2"} Dec 06 16:02:26 crc kubenswrapper[4813]: I1206 16:02:26.216739 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-88qpf" Dec 06 16:02:26 crc kubenswrapper[4813]: I1206 16:02:26.292708 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwpk7\" (UniqueName: \"kubernetes.io/projected/60776264-205f-4541-8d51-3d0d3aec891d-kube-api-access-gwpk7\") pod \"60776264-205f-4541-8d51-3d0d3aec891d\" (UID: \"60776264-205f-4541-8d51-3d0d3aec891d\") " Dec 06 16:02:26 crc kubenswrapper[4813]: I1206 16:02:26.292905 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60776264-205f-4541-8d51-3d0d3aec891d-operator-scripts\") pod \"60776264-205f-4541-8d51-3d0d3aec891d\" (UID: \"60776264-205f-4541-8d51-3d0d3aec891d\") " Dec 06 16:02:26 crc kubenswrapper[4813]: I1206 16:02:26.294171 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60776264-205f-4541-8d51-3d0d3aec891d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "60776264-205f-4541-8d51-3d0d3aec891d" (UID: "60776264-205f-4541-8d51-3d0d3aec891d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:02:26 crc kubenswrapper[4813]: I1206 16:02:26.300215 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60776264-205f-4541-8d51-3d0d3aec891d-kube-api-access-gwpk7" (OuterVolumeSpecName: "kube-api-access-gwpk7") pod "60776264-205f-4541-8d51-3d0d3aec891d" (UID: "60776264-205f-4541-8d51-3d0d3aec891d"). InnerVolumeSpecName "kube-api-access-gwpk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:02:26 crc kubenswrapper[4813]: I1206 16:02:26.356024 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-6nm8x" Dec 06 16:02:26 crc kubenswrapper[4813]: I1206 16:02:26.382453 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4flrx" Dec 06 16:02:26 crc kubenswrapper[4813]: I1206 16:02:26.395117 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwpk7\" (UniqueName: \"kubernetes.io/projected/60776264-205f-4541-8d51-3d0d3aec891d-kube-api-access-gwpk7\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:26 crc kubenswrapper[4813]: I1206 16:02:26.395148 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60776264-205f-4541-8d51-3d0d3aec891d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:26 crc kubenswrapper[4813]: I1206 16:02:26.495996 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/93edeafb-90c8-4ede-b6f4-608b67aadc37-operator-scripts\") pod \"93edeafb-90c8-4ede-b6f4-608b67aadc37\" (UID: \"93edeafb-90c8-4ede-b6f4-608b67aadc37\") " Dec 06 16:02:26 crc kubenswrapper[4813]: I1206 16:02:26.496100 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxjbb\" (UniqueName: \"kubernetes.io/projected/93edeafb-90c8-4ede-b6f4-608b67aadc37-kube-api-access-bxjbb\") pod \"93edeafb-90c8-4ede-b6f4-608b67aadc37\" (UID: \"93edeafb-90c8-4ede-b6f4-608b67aadc37\") " Dec 06 16:02:26 crc kubenswrapper[4813]: I1206 16:02:26.496143 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/235e0c8b-b891-4c0c-9676-d0df7a2b6836-operator-scripts\") pod \"235e0c8b-b891-4c0c-9676-d0df7a2b6836\" (UID: \"235e0c8b-b891-4c0c-9676-d0df7a2b6836\") " Dec 06 16:02:26 crc kubenswrapper[4813]: I1206 16:02:26.496320 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cggz6\" (UniqueName: \"kubernetes.io/projected/235e0c8b-b891-4c0c-9676-d0df7a2b6836-kube-api-access-cggz6\") pod \"235e0c8b-b891-4c0c-9676-d0df7a2b6836\" (UID: \"235e0c8b-b891-4c0c-9676-d0df7a2b6836\") " Dec 06 16:02:26 crc kubenswrapper[4813]: I1206 16:02:26.497521 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/235e0c8b-b891-4c0c-9676-d0df7a2b6836-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "235e0c8b-b891-4c0c-9676-d0df7a2b6836" (UID: "235e0c8b-b891-4c0c-9676-d0df7a2b6836"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:02:26 crc kubenswrapper[4813]: I1206 16:02:26.497789 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93edeafb-90c8-4ede-b6f4-608b67aadc37-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "93edeafb-90c8-4ede-b6f4-608b67aadc37" (UID: "93edeafb-90c8-4ede-b6f4-608b67aadc37"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:02:26 crc kubenswrapper[4813]: I1206 16:02:26.500058 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93edeafb-90c8-4ede-b6f4-608b67aadc37-kube-api-access-bxjbb" (OuterVolumeSpecName: "kube-api-access-bxjbb") pod "93edeafb-90c8-4ede-b6f4-608b67aadc37" (UID: "93edeafb-90c8-4ede-b6f4-608b67aadc37"). InnerVolumeSpecName "kube-api-access-bxjbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:02:26 crc kubenswrapper[4813]: I1206 16:02:26.500493 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/235e0c8b-b891-4c0c-9676-d0df7a2b6836-kube-api-access-cggz6" (OuterVolumeSpecName: "kube-api-access-cggz6") pod "235e0c8b-b891-4c0c-9676-d0df7a2b6836" (UID: "235e0c8b-b891-4c0c-9676-d0df7a2b6836"). InnerVolumeSpecName "kube-api-access-cggz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:02:26 crc kubenswrapper[4813]: I1206 16:02:26.598151 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/93edeafb-90c8-4ede-b6f4-608b67aadc37-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:26 crc kubenswrapper[4813]: I1206 16:02:26.598180 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxjbb\" (UniqueName: \"kubernetes.io/projected/93edeafb-90c8-4ede-b6f4-608b67aadc37-kube-api-access-bxjbb\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:26 crc kubenswrapper[4813]: I1206 16:02:26.598191 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/235e0c8b-b891-4c0c-9676-d0df7a2b6836-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:26 crc kubenswrapper[4813]: I1206 16:02:26.598201 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cggz6\" (UniqueName: \"kubernetes.io/projected/235e0c8b-b891-4c0c-9676-d0df7a2b6836-kube-api-access-cggz6\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:26 crc kubenswrapper[4813]: I1206 16:02:26.692193 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-6nm8x" event={"ID":"93edeafb-90c8-4ede-b6f4-608b67aadc37","Type":"ContainerDied","Data":"5833b8ecfedb13fb375d204948cf8b060a974af2ecf6223fcbbefe0dd9679c42"} Dec 06 16:02:26 crc kubenswrapper[4813]: I1206 16:02:26.692229 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5833b8ecfedb13fb375d204948cf8b060a974af2ecf6223fcbbefe0dd9679c42" Dec 06 16:02:26 crc kubenswrapper[4813]: I1206 16:02:26.692301 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-6nm8x" Dec 06 16:02:26 crc kubenswrapper[4813]: I1206 16:02:26.700840 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4flrx" event={"ID":"235e0c8b-b891-4c0c-9676-d0df7a2b6836","Type":"ContainerDied","Data":"f8f895887285a01de2487b83a3824161b7731aba899f19be687c15b051cb155c"} Dec 06 16:02:26 crc kubenswrapper[4813]: I1206 16:02:26.700882 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8f895887285a01de2487b83a3824161b7731aba899f19be687c15b051cb155c" Dec 06 16:02:26 crc kubenswrapper[4813]: I1206 16:02:26.700883 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4flrx" Dec 06 16:02:26 crc kubenswrapper[4813]: I1206 16:02:26.712239 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-88qpf" event={"ID":"60776264-205f-4541-8d51-3d0d3aec891d","Type":"ContainerDied","Data":"a4489a86a2e5c777635390646f071883caa0088ddd0c0d98c90d7b12d1072339"} Dec 06 16:02:26 crc kubenswrapper[4813]: I1206 16:02:26.712292 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4489a86a2e5c777635390646f071883caa0088ddd0c0d98c90d7b12d1072339" Dec 06 16:02:26 crc kubenswrapper[4813]: I1206 16:02:26.712321 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-88qpf" Dec 06 16:02:27 crc kubenswrapper[4813]: I1206 16:02:27.719994 4813 generic.go:334] "Generic (PLEG): container finished" podID="27b4a8bf-5c37-4936-8ced-6a1d02249dd1" containerID="1ef292affd710b39999c751c2f5a2ca32230e936cfa573d521c6d139e9f27c7c" exitCode=0 Dec 06 16:02:27 crc kubenswrapper[4813]: I1206 16:02:27.720176 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tzd8d" event={"ID":"27b4a8bf-5c37-4936-8ced-6a1d02249dd1","Type":"ContainerDied","Data":"1ef292affd710b39999c751c2f5a2ca32230e936cfa573d521c6d139e9f27c7c"} Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.103443 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-2f15-account-create-update-2gvw4" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.129551 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c831-account-create-update-h8lzr" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.155363 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-24e9-account-create-update-ht2lg" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.160724 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/612d9b68-b55f-40f5-a45a-c1d7a3b32f2b-operator-scripts\") pod \"612d9b68-b55f-40f5-a45a-c1d7a3b32f2b\" (UID: \"612d9b68-b55f-40f5-a45a-c1d7a3b32f2b\") " Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.160863 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vn6jh\" (UniqueName: \"kubernetes.io/projected/612d9b68-b55f-40f5-a45a-c1d7a3b32f2b-kube-api-access-vn6jh\") pod \"612d9b68-b55f-40f5-a45a-c1d7a3b32f2b\" (UID: \"612d9b68-b55f-40f5-a45a-c1d7a3b32f2b\") " Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.162514 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/612d9b68-b55f-40f5-a45a-c1d7a3b32f2b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "612d9b68-b55f-40f5-a45a-c1d7a3b32f2b" (UID: "612d9b68-b55f-40f5-a45a-c1d7a3b32f2b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.165818 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/612d9b68-b55f-40f5-a45a-c1d7a3b32f2b-kube-api-access-vn6jh" (OuterVolumeSpecName: "kube-api-access-vn6jh") pod "612d9b68-b55f-40f5-a45a-c1d7a3b32f2b" (UID: "612d9b68-b55f-40f5-a45a-c1d7a3b32f2b"). InnerVolumeSpecName "kube-api-access-vn6jh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.172010 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tzd8d" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.262791 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khnfv\" (UniqueName: \"kubernetes.io/projected/3579ce45-95fa-44af-a164-3153a1c873ac-kube-api-access-khnfv\") pod \"3579ce45-95fa-44af-a164-3153a1c873ac\" (UID: \"3579ce45-95fa-44af-a164-3153a1c873ac\") " Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.262885 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5xhm\" (UniqueName: \"kubernetes.io/projected/27b4a8bf-5c37-4936-8ced-6a1d02249dd1-kube-api-access-f5xhm\") pod \"27b4a8bf-5c37-4936-8ced-6a1d02249dd1\" (UID: \"27b4a8bf-5c37-4936-8ced-6a1d02249dd1\") " Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.262919 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27b4a8bf-5c37-4936-8ced-6a1d02249dd1-combined-ca-bundle\") pod \"27b4a8bf-5c37-4936-8ced-6a1d02249dd1\" (UID: \"27b4a8bf-5c37-4936-8ced-6a1d02249dd1\") " Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.262986 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/27b4a8bf-5c37-4936-8ced-6a1d02249dd1-db-sync-config-data\") pod \"27b4a8bf-5c37-4936-8ced-6a1d02249dd1\" (UID: \"27b4a8bf-5c37-4936-8ced-6a1d02249dd1\") " Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.263013 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27b4a8bf-5c37-4936-8ced-6a1d02249dd1-config-data\") pod \"27b4a8bf-5c37-4936-8ced-6a1d02249dd1\" (UID: \"27b4a8bf-5c37-4936-8ced-6a1d02249dd1\") " Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.263038 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3579ce45-95fa-44af-a164-3153a1c873ac-operator-scripts\") pod \"3579ce45-95fa-44af-a164-3153a1c873ac\" (UID: \"3579ce45-95fa-44af-a164-3153a1c873ac\") " Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.263099 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9-operator-scripts\") pod \"9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9\" (UID: \"9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9\") " Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.263185 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbp8s\" (UniqueName: \"kubernetes.io/projected/9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9-kube-api-access-nbp8s\") pod \"9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9\" (UID: \"9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9\") " Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.263834 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3579ce45-95fa-44af-a164-3153a1c873ac-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3579ce45-95fa-44af-a164-3153a1c873ac" (UID: "3579ce45-95fa-44af-a164-3153a1c873ac"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.264119 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9" (UID: "9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.265672 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3579ce45-95fa-44af-a164-3153a1c873ac-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.265699 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/612d9b68-b55f-40f5-a45a-c1d7a3b32f2b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.265711 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.265722 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vn6jh\" (UniqueName: \"kubernetes.io/projected/612d9b68-b55f-40f5-a45a-c1d7a3b32f2b-kube-api-access-vn6jh\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.267836 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27b4a8bf-5c37-4936-8ced-6a1d02249dd1-kube-api-access-f5xhm" (OuterVolumeSpecName: "kube-api-access-f5xhm") pod "27b4a8bf-5c37-4936-8ced-6a1d02249dd1" (UID: "27b4a8bf-5c37-4936-8ced-6a1d02249dd1"). InnerVolumeSpecName "kube-api-access-f5xhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.268127 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3579ce45-95fa-44af-a164-3153a1c873ac-kube-api-access-khnfv" (OuterVolumeSpecName: "kube-api-access-khnfv") pod "3579ce45-95fa-44af-a164-3153a1c873ac" (UID: "3579ce45-95fa-44af-a164-3153a1c873ac"). InnerVolumeSpecName "kube-api-access-khnfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.268906 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9-kube-api-access-nbp8s" (OuterVolumeSpecName: "kube-api-access-nbp8s") pod "9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9" (UID: "9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9"). InnerVolumeSpecName "kube-api-access-nbp8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.269601 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27b4a8bf-5c37-4936-8ced-6a1d02249dd1-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "27b4a8bf-5c37-4936-8ced-6a1d02249dd1" (UID: "27b4a8bf-5c37-4936-8ced-6a1d02249dd1"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.289799 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27b4a8bf-5c37-4936-8ced-6a1d02249dd1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "27b4a8bf-5c37-4936-8ced-6a1d02249dd1" (UID: "27b4a8bf-5c37-4936-8ced-6a1d02249dd1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.293446 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.334070 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27b4a8bf-5c37-4936-8ced-6a1d02249dd1-config-data" (OuterVolumeSpecName: "config-data") pod "27b4a8bf-5c37-4936-8ced-6a1d02249dd1" (UID: "27b4a8bf-5c37-4936-8ced-6a1d02249dd1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.369220 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbp8s\" (UniqueName: \"kubernetes.io/projected/9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9-kube-api-access-nbp8s\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.369252 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khnfv\" (UniqueName: \"kubernetes.io/projected/3579ce45-95fa-44af-a164-3153a1c873ac-kube-api-access-khnfv\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.369314 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5xhm\" (UniqueName: \"kubernetes.io/projected/27b4a8bf-5c37-4936-8ced-6a1d02249dd1-kube-api-access-f5xhm\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.369324 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27b4a8bf-5c37-4936-8ced-6a1d02249dd1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.369332 4813 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/27b4a8bf-5c37-4936-8ced-6a1d02249dd1-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.369340 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27b4a8bf-5c37-4936-8ced-6a1d02249dd1-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.381925 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-8l7jf"] Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.382187 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-8l7jf" podUID="b4400510-3d90-46c6-ba53-d964f310719a" containerName="dnsmasq-dns" containerID="cri-o://11fcb034dbd10d0bf3b59810737248189753f8ccf7472663584df920cd838432" gracePeriod=10 Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.758197 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-8l7jf" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.786086 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tzd8d" event={"ID":"27b4a8bf-5c37-4936-8ced-6a1d02249dd1","Type":"ContainerDied","Data":"291932d28876c2271569b42c8c1d40cf9cf7da40dc6dd77155b98e1840d90b48"} Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.786126 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="291932d28876c2271569b42c8c1d40cf9cf7da40dc6dd77155b98e1840d90b48" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.786185 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tzd8d" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.792097 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c831-account-create-update-h8lzr" event={"ID":"3579ce45-95fa-44af-a164-3153a1c873ac","Type":"ContainerDied","Data":"9902d3fcf6cfcaf401d10999ba5f1dcc1b49e565991c54b9b050bc8dce029ded"} Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.792132 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9902d3fcf6cfcaf401d10999ba5f1dcc1b49e565991c54b9b050bc8dce029ded" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.792191 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c831-account-create-update-h8lzr" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.795585 4813 generic.go:334] "Generic (PLEG): container finished" podID="b4400510-3d90-46c6-ba53-d964f310719a" containerID="11fcb034dbd10d0bf3b59810737248189753f8ccf7472663584df920cd838432" exitCode=0 Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.795637 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-8l7jf" event={"ID":"b4400510-3d90-46c6-ba53-d964f310719a","Type":"ContainerDied","Data":"11fcb034dbd10d0bf3b59810737248189753f8ccf7472663584df920cd838432"} Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.795662 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-8l7jf" event={"ID":"b4400510-3d90-46c6-ba53-d964f310719a","Type":"ContainerDied","Data":"20ac4a2474e39665abda47620de8302af53ef196d3bc48519ff532816e47db72"} Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.795678 4813 scope.go:117] "RemoveContainer" containerID="11fcb034dbd10d0bf3b59810737248189753f8ccf7472663584df920cd838432" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.795769 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-8l7jf" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.799843 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-24e9-account-create-update-ht2lg" event={"ID":"9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9","Type":"ContainerDied","Data":"6a968d37f1afc4a031f2ca5c62b38b29d9e027398e9f66f0b3fef38ca3257560"} Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.799872 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a968d37f1afc4a031f2ca5c62b38b29d9e027398e9f66f0b3fef38ca3257560" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.799918 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-24e9-account-create-update-ht2lg" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.807086 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-2f15-account-create-update-2gvw4" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.807336 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-2f15-account-create-update-2gvw4" event={"ID":"612d9b68-b55f-40f5-a45a-c1d7a3b32f2b","Type":"ContainerDied","Data":"cffcc9065a4a92e0bce0108f3aff0758ec64a8aa52e73efdadbe9b5cd16dcede"} Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.807370 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cffcc9065a4a92e0bce0108f3aff0758ec64a8aa52e73efdadbe9b5cd16dcede" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.810240 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jz5zm" event={"ID":"4cc7028f-2019-4d85-84d9-7e7bb05216ec","Type":"ContainerStarted","Data":"612896523b7d109cffa9908510c0d98cc71cce4af36e28f74d2c8d76282ad269"} Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.832176 4813 scope.go:117] "RemoveContainer" containerID="bd8496a393a15d54119a5e6ab856cd1a41772d40d077496f866f1d392899bf4b" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.844970 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-jz5zm" podStartSLOduration=3.18751087 podStartE2EDuration="8.844947705s" podCreationTimestamp="2025-12-06 16:02:22 +0000 UTC" firstStartedPulling="2025-12-06 16:02:24.279757882 +0000 UTC m=+984.170637458" lastFinishedPulling="2025-12-06 16:02:29.937194727 +0000 UTC m=+989.828074293" observedRunningTime="2025-12-06 16:02:30.833950014 +0000 UTC m=+990.724829610" watchObservedRunningTime="2025-12-06 16:02:30.844947705 +0000 UTC m=+990.735827281" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.854274 4813 scope.go:117] "RemoveContainer" containerID="11fcb034dbd10d0bf3b59810737248189753f8ccf7472663584df920cd838432" Dec 06 16:02:30 crc kubenswrapper[4813]: E1206 16:02:30.855213 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11fcb034dbd10d0bf3b59810737248189753f8ccf7472663584df920cd838432\": container with ID starting with 11fcb034dbd10d0bf3b59810737248189753f8ccf7472663584df920cd838432 not found: ID does not exist" containerID="11fcb034dbd10d0bf3b59810737248189753f8ccf7472663584df920cd838432" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.855310 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11fcb034dbd10d0bf3b59810737248189753f8ccf7472663584df920cd838432"} err="failed to get container status \"11fcb034dbd10d0bf3b59810737248189753f8ccf7472663584df920cd838432\": rpc error: code = NotFound desc = could not find container \"11fcb034dbd10d0bf3b59810737248189753f8ccf7472663584df920cd838432\": container with ID starting with 11fcb034dbd10d0bf3b59810737248189753f8ccf7472663584df920cd838432 not found: ID does not exist" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.855336 4813 scope.go:117] "RemoveContainer" containerID="bd8496a393a15d54119a5e6ab856cd1a41772d40d077496f866f1d392899bf4b" Dec 06 16:02:30 crc kubenswrapper[4813]: E1206 16:02:30.856927 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd8496a393a15d54119a5e6ab856cd1a41772d40d077496f866f1d392899bf4b\": container with ID starting with bd8496a393a15d54119a5e6ab856cd1a41772d40d077496f866f1d392899bf4b not found: ID does not exist" containerID="bd8496a393a15d54119a5e6ab856cd1a41772d40d077496f866f1d392899bf4b" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.856977 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd8496a393a15d54119a5e6ab856cd1a41772d40d077496f866f1d392899bf4b"} err="failed to get container status \"bd8496a393a15d54119a5e6ab856cd1a41772d40d077496f866f1d392899bf4b\": rpc error: code = NotFound desc = could not find container \"bd8496a393a15d54119a5e6ab856cd1a41772d40d077496f866f1d392899bf4b\": container with ID starting with bd8496a393a15d54119a5e6ab856cd1a41772d40d077496f866f1d392899bf4b not found: ID does not exist" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.881536 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4400510-3d90-46c6-ba53-d964f310719a-ovsdbserver-nb\") pod \"b4400510-3d90-46c6-ba53-d964f310719a\" (UID: \"b4400510-3d90-46c6-ba53-d964f310719a\") " Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.881611 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4400510-3d90-46c6-ba53-d964f310719a-dns-svc\") pod \"b4400510-3d90-46c6-ba53-d964f310719a\" (UID: \"b4400510-3d90-46c6-ba53-d964f310719a\") " Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.883497 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4400510-3d90-46c6-ba53-d964f310719a-config\") pod \"b4400510-3d90-46c6-ba53-d964f310719a\" (UID: \"b4400510-3d90-46c6-ba53-d964f310719a\") " Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.884007 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4400510-3d90-46c6-ba53-d964f310719a-ovsdbserver-sb\") pod \"b4400510-3d90-46c6-ba53-d964f310719a\" (UID: \"b4400510-3d90-46c6-ba53-d964f310719a\") " Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.885657 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2wzt\" (UniqueName: \"kubernetes.io/projected/b4400510-3d90-46c6-ba53-d964f310719a-kube-api-access-x2wzt\") pod \"b4400510-3d90-46c6-ba53-d964f310719a\" (UID: \"b4400510-3d90-46c6-ba53-d964f310719a\") " Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.896974 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4400510-3d90-46c6-ba53-d964f310719a-kube-api-access-x2wzt" (OuterVolumeSpecName: "kube-api-access-x2wzt") pod "b4400510-3d90-46c6-ba53-d964f310719a" (UID: "b4400510-3d90-46c6-ba53-d964f310719a"). InnerVolumeSpecName "kube-api-access-x2wzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.924188 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4400510-3d90-46c6-ba53-d964f310719a-config" (OuterVolumeSpecName: "config") pod "b4400510-3d90-46c6-ba53-d964f310719a" (UID: "b4400510-3d90-46c6-ba53-d964f310719a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.931741 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4400510-3d90-46c6-ba53-d964f310719a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b4400510-3d90-46c6-ba53-d964f310719a" (UID: "b4400510-3d90-46c6-ba53-d964f310719a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.936382 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4400510-3d90-46c6-ba53-d964f310719a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b4400510-3d90-46c6-ba53-d964f310719a" (UID: "b4400510-3d90-46c6-ba53-d964f310719a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.946367 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4400510-3d90-46c6-ba53-d964f310719a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b4400510-3d90-46c6-ba53-d964f310719a" (UID: "b4400510-3d90-46c6-ba53-d964f310719a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.987823 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4400510-3d90-46c6-ba53-d964f310719a-config\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.987856 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4400510-3d90-46c6-ba53-d964f310719a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.987867 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2wzt\" (UniqueName: \"kubernetes.io/projected/b4400510-3d90-46c6-ba53-d964f310719a-kube-api-access-x2wzt\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.987877 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4400510-3d90-46c6-ba53-d964f310719a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:30 crc kubenswrapper[4813]: I1206 16:02:30.987886 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4400510-3d90-46c6-ba53-d964f310719a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.130802 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-8l7jf"] Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.141717 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-8l7jf"] Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.492798 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-gtr8j"] Dec 06 16:02:31 crc kubenswrapper[4813]: E1206 16:02:31.493119 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4400510-3d90-46c6-ba53-d964f310719a" containerName="init" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.493136 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4400510-3d90-46c6-ba53-d964f310719a" containerName="init" Dec 06 16:02:31 crc kubenswrapper[4813]: E1206 16:02:31.493150 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="235e0c8b-b891-4c0c-9676-d0df7a2b6836" containerName="mariadb-database-create" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.493157 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="235e0c8b-b891-4c0c-9676-d0df7a2b6836" containerName="mariadb-database-create" Dec 06 16:02:31 crc kubenswrapper[4813]: E1206 16:02:31.493166 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93edeafb-90c8-4ede-b6f4-608b67aadc37" containerName="mariadb-database-create" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.493173 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="93edeafb-90c8-4ede-b6f4-608b67aadc37" containerName="mariadb-database-create" Dec 06 16:02:31 crc kubenswrapper[4813]: E1206 16:02:31.493186 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="612d9b68-b55f-40f5-a45a-c1d7a3b32f2b" containerName="mariadb-account-create-update" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.493192 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="612d9b68-b55f-40f5-a45a-c1d7a3b32f2b" containerName="mariadb-account-create-update" Dec 06 16:02:31 crc kubenswrapper[4813]: E1206 16:02:31.493213 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3579ce45-95fa-44af-a164-3153a1c873ac" containerName="mariadb-account-create-update" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.493218 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="3579ce45-95fa-44af-a164-3153a1c873ac" containerName="mariadb-account-create-update" Dec 06 16:02:31 crc kubenswrapper[4813]: E1206 16:02:31.493226 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27b4a8bf-5c37-4936-8ced-6a1d02249dd1" containerName="glance-db-sync" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.493232 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="27b4a8bf-5c37-4936-8ced-6a1d02249dd1" containerName="glance-db-sync" Dec 06 16:02:31 crc kubenswrapper[4813]: E1206 16:02:31.493243 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60776264-205f-4541-8d51-3d0d3aec891d" containerName="mariadb-database-create" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.493249 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="60776264-205f-4541-8d51-3d0d3aec891d" containerName="mariadb-database-create" Dec 06 16:02:31 crc kubenswrapper[4813]: E1206 16:02:31.493271 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9" containerName="mariadb-account-create-update" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.493277 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9" containerName="mariadb-account-create-update" Dec 06 16:02:31 crc kubenswrapper[4813]: E1206 16:02:31.493289 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4400510-3d90-46c6-ba53-d964f310719a" containerName="dnsmasq-dns" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.493295 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4400510-3d90-46c6-ba53-d964f310719a" containerName="dnsmasq-dns" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.493443 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="60776264-205f-4541-8d51-3d0d3aec891d" containerName="mariadb-database-create" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.493459 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="27b4a8bf-5c37-4936-8ced-6a1d02249dd1" containerName="glance-db-sync" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.493469 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="235e0c8b-b891-4c0c-9676-d0df7a2b6836" containerName="mariadb-database-create" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.493481 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9" containerName="mariadb-account-create-update" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.493491 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4400510-3d90-46c6-ba53-d964f310719a" containerName="dnsmasq-dns" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.493505 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="93edeafb-90c8-4ede-b6f4-608b67aadc37" containerName="mariadb-database-create" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.493522 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="3579ce45-95fa-44af-a164-3153a1c873ac" containerName="mariadb-account-create-update" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.493531 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="612d9b68-b55f-40f5-a45a-c1d7a3b32f2b" containerName="mariadb-account-create-update" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.494314 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-gtr8j" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.515240 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-gtr8j"] Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.597058 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htp6w\" (UniqueName: \"kubernetes.io/projected/d1f29fa9-6e22-45be-97c4-4317914c1aa6-kube-api-access-htp6w\") pod \"dnsmasq-dns-5f59b8f679-gtr8j\" (UID: \"d1f29fa9-6e22-45be-97c4-4317914c1aa6\") " pod="openstack/dnsmasq-dns-5f59b8f679-gtr8j" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.597114 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1f29fa9-6e22-45be-97c4-4317914c1aa6-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-gtr8j\" (UID: \"d1f29fa9-6e22-45be-97c4-4317914c1aa6\") " pod="openstack/dnsmasq-dns-5f59b8f679-gtr8j" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.597155 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1f29fa9-6e22-45be-97c4-4317914c1aa6-config\") pod \"dnsmasq-dns-5f59b8f679-gtr8j\" (UID: \"d1f29fa9-6e22-45be-97c4-4317914c1aa6\") " pod="openstack/dnsmasq-dns-5f59b8f679-gtr8j" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.597233 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d1f29fa9-6e22-45be-97c4-4317914c1aa6-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-gtr8j\" (UID: \"d1f29fa9-6e22-45be-97c4-4317914c1aa6\") " pod="openstack/dnsmasq-dns-5f59b8f679-gtr8j" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.597287 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d1f29fa9-6e22-45be-97c4-4317914c1aa6-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-gtr8j\" (UID: \"d1f29fa9-6e22-45be-97c4-4317914c1aa6\") " pod="openstack/dnsmasq-dns-5f59b8f679-gtr8j" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.597330 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d1f29fa9-6e22-45be-97c4-4317914c1aa6-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-gtr8j\" (UID: \"d1f29fa9-6e22-45be-97c4-4317914c1aa6\") " pod="openstack/dnsmasq-dns-5f59b8f679-gtr8j" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.698855 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1f29fa9-6e22-45be-97c4-4317914c1aa6-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-gtr8j\" (UID: \"d1f29fa9-6e22-45be-97c4-4317914c1aa6\") " pod="openstack/dnsmasq-dns-5f59b8f679-gtr8j" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.698919 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1f29fa9-6e22-45be-97c4-4317914c1aa6-config\") pod \"dnsmasq-dns-5f59b8f679-gtr8j\" (UID: \"d1f29fa9-6e22-45be-97c4-4317914c1aa6\") " pod="openstack/dnsmasq-dns-5f59b8f679-gtr8j" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.698959 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d1f29fa9-6e22-45be-97c4-4317914c1aa6-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-gtr8j\" (UID: \"d1f29fa9-6e22-45be-97c4-4317914c1aa6\") " pod="openstack/dnsmasq-dns-5f59b8f679-gtr8j" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.698986 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d1f29fa9-6e22-45be-97c4-4317914c1aa6-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-gtr8j\" (UID: \"d1f29fa9-6e22-45be-97c4-4317914c1aa6\") " pod="openstack/dnsmasq-dns-5f59b8f679-gtr8j" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.699010 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d1f29fa9-6e22-45be-97c4-4317914c1aa6-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-gtr8j\" (UID: \"d1f29fa9-6e22-45be-97c4-4317914c1aa6\") " pod="openstack/dnsmasq-dns-5f59b8f679-gtr8j" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.699052 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htp6w\" (UniqueName: \"kubernetes.io/projected/d1f29fa9-6e22-45be-97c4-4317914c1aa6-kube-api-access-htp6w\") pod \"dnsmasq-dns-5f59b8f679-gtr8j\" (UID: \"d1f29fa9-6e22-45be-97c4-4317914c1aa6\") " pod="openstack/dnsmasq-dns-5f59b8f679-gtr8j" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.699636 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1f29fa9-6e22-45be-97c4-4317914c1aa6-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-gtr8j\" (UID: \"d1f29fa9-6e22-45be-97c4-4317914c1aa6\") " pod="openstack/dnsmasq-dns-5f59b8f679-gtr8j" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.700197 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d1f29fa9-6e22-45be-97c4-4317914c1aa6-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-gtr8j\" (UID: \"d1f29fa9-6e22-45be-97c4-4317914c1aa6\") " pod="openstack/dnsmasq-dns-5f59b8f679-gtr8j" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.700227 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d1f29fa9-6e22-45be-97c4-4317914c1aa6-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-gtr8j\" (UID: \"d1f29fa9-6e22-45be-97c4-4317914c1aa6\") " pod="openstack/dnsmasq-dns-5f59b8f679-gtr8j" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.700343 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d1f29fa9-6e22-45be-97c4-4317914c1aa6-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-gtr8j\" (UID: \"d1f29fa9-6e22-45be-97c4-4317914c1aa6\") " pod="openstack/dnsmasq-dns-5f59b8f679-gtr8j" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.700744 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1f29fa9-6e22-45be-97c4-4317914c1aa6-config\") pod \"dnsmasq-dns-5f59b8f679-gtr8j\" (UID: \"d1f29fa9-6e22-45be-97c4-4317914c1aa6\") " pod="openstack/dnsmasq-dns-5f59b8f679-gtr8j" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.730513 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htp6w\" (UniqueName: \"kubernetes.io/projected/d1f29fa9-6e22-45be-97c4-4317914c1aa6-kube-api-access-htp6w\") pod \"dnsmasq-dns-5f59b8f679-gtr8j\" (UID: \"d1f29fa9-6e22-45be-97c4-4317914c1aa6\") " pod="openstack/dnsmasq-dns-5f59b8f679-gtr8j" Dec 06 16:02:31 crc kubenswrapper[4813]: I1206 16:02:31.809462 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-gtr8j" Dec 06 16:02:32 crc kubenswrapper[4813]: I1206 16:02:32.259438 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-gtr8j"] Dec 06 16:02:32 crc kubenswrapper[4813]: W1206 16:02:32.263884 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1f29fa9_6e22_45be_97c4_4317914c1aa6.slice/crio-8d45f528928bcc87089be5eb4ef976ce4f2a8ea362b0dacc88d08611e4cb3fbd WatchSource:0}: Error finding container 8d45f528928bcc87089be5eb4ef976ce4f2a8ea362b0dacc88d08611e4cb3fbd: Status 404 returned error can't find the container with id 8d45f528928bcc87089be5eb4ef976ce4f2a8ea362b0dacc88d08611e4cb3fbd Dec 06 16:02:32 crc kubenswrapper[4813]: I1206 16:02:32.495482 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4400510-3d90-46c6-ba53-d964f310719a" path="/var/lib/kubelet/pods/b4400510-3d90-46c6-ba53-d964f310719a/volumes" Dec 06 16:02:32 crc kubenswrapper[4813]: I1206 16:02:32.828610 4813 generic.go:334] "Generic (PLEG): container finished" podID="d1f29fa9-6e22-45be-97c4-4317914c1aa6" containerID="95d42e55f4236eeb7fa3f110922fb389ad84293e533194f9070db39b74b0e7e9" exitCode=0 Dec 06 16:02:32 crc kubenswrapper[4813]: I1206 16:02:32.828654 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-gtr8j" event={"ID":"d1f29fa9-6e22-45be-97c4-4317914c1aa6","Type":"ContainerDied","Data":"95d42e55f4236eeb7fa3f110922fb389ad84293e533194f9070db39b74b0e7e9"} Dec 06 16:02:32 crc kubenswrapper[4813]: I1206 16:02:32.828707 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-gtr8j" event={"ID":"d1f29fa9-6e22-45be-97c4-4317914c1aa6","Type":"ContainerStarted","Data":"8d45f528928bcc87089be5eb4ef976ce4f2a8ea362b0dacc88d08611e4cb3fbd"} Dec 06 16:02:33 crc kubenswrapper[4813]: I1206 16:02:33.837082 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-gtr8j" event={"ID":"d1f29fa9-6e22-45be-97c4-4317914c1aa6","Type":"ContainerStarted","Data":"6e91d9bd84b543fe730a650e2cb0ad54ef7b2740d8aff51308221066d025c3cc"} Dec 06 16:02:33 crc kubenswrapper[4813]: I1206 16:02:33.837664 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f59b8f679-gtr8j" Dec 06 16:02:33 crc kubenswrapper[4813]: I1206 16:02:33.838556 4813 generic.go:334] "Generic (PLEG): container finished" podID="4cc7028f-2019-4d85-84d9-7e7bb05216ec" containerID="612896523b7d109cffa9908510c0d98cc71cce4af36e28f74d2c8d76282ad269" exitCode=0 Dec 06 16:02:33 crc kubenswrapper[4813]: I1206 16:02:33.838607 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jz5zm" event={"ID":"4cc7028f-2019-4d85-84d9-7e7bb05216ec","Type":"ContainerDied","Data":"612896523b7d109cffa9908510c0d98cc71cce4af36e28f74d2c8d76282ad269"} Dec 06 16:02:33 crc kubenswrapper[4813]: I1206 16:02:33.856960 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f59b8f679-gtr8j" podStartSLOduration=2.856945568 podStartE2EDuration="2.856945568s" podCreationTimestamp="2025-12-06 16:02:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:02:33.853486557 +0000 UTC m=+993.744366133" watchObservedRunningTime="2025-12-06 16:02:33.856945568 +0000 UTC m=+993.747825144" Dec 06 16:02:35 crc kubenswrapper[4813]: I1206 16:02:35.202855 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jz5zm" Dec 06 16:02:35 crc kubenswrapper[4813]: I1206 16:02:35.259274 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cc7028f-2019-4d85-84d9-7e7bb05216ec-combined-ca-bundle\") pod \"4cc7028f-2019-4d85-84d9-7e7bb05216ec\" (UID: \"4cc7028f-2019-4d85-84d9-7e7bb05216ec\") " Dec 06 16:02:35 crc kubenswrapper[4813]: I1206 16:02:35.259319 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rq4zm\" (UniqueName: \"kubernetes.io/projected/4cc7028f-2019-4d85-84d9-7e7bb05216ec-kube-api-access-rq4zm\") pod \"4cc7028f-2019-4d85-84d9-7e7bb05216ec\" (UID: \"4cc7028f-2019-4d85-84d9-7e7bb05216ec\") " Dec 06 16:02:35 crc kubenswrapper[4813]: I1206 16:02:35.259352 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4cc7028f-2019-4d85-84d9-7e7bb05216ec-config-data\") pod \"4cc7028f-2019-4d85-84d9-7e7bb05216ec\" (UID: \"4cc7028f-2019-4d85-84d9-7e7bb05216ec\") " Dec 06 16:02:35 crc kubenswrapper[4813]: I1206 16:02:35.265295 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cc7028f-2019-4d85-84d9-7e7bb05216ec-kube-api-access-rq4zm" (OuterVolumeSpecName: "kube-api-access-rq4zm") pod "4cc7028f-2019-4d85-84d9-7e7bb05216ec" (UID: "4cc7028f-2019-4d85-84d9-7e7bb05216ec"). InnerVolumeSpecName "kube-api-access-rq4zm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:02:35 crc kubenswrapper[4813]: I1206 16:02:35.286996 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4cc7028f-2019-4d85-84d9-7e7bb05216ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4cc7028f-2019-4d85-84d9-7e7bb05216ec" (UID: "4cc7028f-2019-4d85-84d9-7e7bb05216ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:02:35 crc kubenswrapper[4813]: I1206 16:02:35.312770 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4cc7028f-2019-4d85-84d9-7e7bb05216ec-config-data" (OuterVolumeSpecName: "config-data") pod "4cc7028f-2019-4d85-84d9-7e7bb05216ec" (UID: "4cc7028f-2019-4d85-84d9-7e7bb05216ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:02:35 crc kubenswrapper[4813]: I1206 16:02:35.360743 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4cc7028f-2019-4d85-84d9-7e7bb05216ec-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:35 crc kubenswrapper[4813]: I1206 16:02:35.360779 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cc7028f-2019-4d85-84d9-7e7bb05216ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:35 crc kubenswrapper[4813]: I1206 16:02:35.360793 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rq4zm\" (UniqueName: \"kubernetes.io/projected/4cc7028f-2019-4d85-84d9-7e7bb05216ec-kube-api-access-rq4zm\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:35 crc kubenswrapper[4813]: I1206 16:02:35.854423 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jz5zm" event={"ID":"4cc7028f-2019-4d85-84d9-7e7bb05216ec","Type":"ContainerDied","Data":"86e889e88c2aa449fc20ce62cef21ab9dcae0677a68ce8fcca2b63445d618a1d"} Dec 06 16:02:35 crc kubenswrapper[4813]: I1206 16:02:35.854465 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="86e889e88c2aa449fc20ce62cef21ab9dcae0677a68ce8fcca2b63445d618a1d" Dec 06 16:02:35 crc kubenswrapper[4813]: I1206 16:02:35.854502 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jz5zm" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.040879 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-gtr8j"] Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.041089 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f59b8f679-gtr8j" podUID="d1f29fa9-6e22-45be-97c4-4317914c1aa6" containerName="dnsmasq-dns" containerID="cri-o://6e91d9bd84b543fe730a650e2cb0ad54ef7b2740d8aff51308221066d025c3cc" gracePeriod=10 Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.089871 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-wrj6z"] Dec 06 16:02:36 crc kubenswrapper[4813]: E1206 16:02:36.090236 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cc7028f-2019-4d85-84d9-7e7bb05216ec" containerName="keystone-db-sync" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.090251 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cc7028f-2019-4d85-84d9-7e7bb05216ec" containerName="keystone-db-sync" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.090422 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cc7028f-2019-4d85-84d9-7e7bb05216ec" containerName="keystone-db-sync" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.093186 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-wrj6z" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.097350 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-zbvhf"] Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.098254 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zbvhf" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.102453 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.102716 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fmvk2" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.103072 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.103320 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.105484 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.134730 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-wrj6z"] Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.146062 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-zbvhf"] Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.172808 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsrk8\" (UniqueName: \"kubernetes.io/projected/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-kube-api-access-gsrk8\") pod \"keystone-bootstrap-zbvhf\" (UID: \"e5d43f84-cbc2-43a7-877c-56d53b8de4b3\") " pod="openstack/keystone-bootstrap-zbvhf" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.172857 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-config-data\") pod \"keystone-bootstrap-zbvhf\" (UID: \"e5d43f84-cbc2-43a7-877c-56d53b8de4b3\") " pod="openstack/keystone-bootstrap-zbvhf" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.172875 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-fernet-keys\") pod \"keystone-bootstrap-zbvhf\" (UID: \"e5d43f84-cbc2-43a7-877c-56d53b8de4b3\") " pod="openstack/keystone-bootstrap-zbvhf" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.172895 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-scripts\") pod \"keystone-bootstrap-zbvhf\" (UID: \"e5d43f84-cbc2-43a7-877c-56d53b8de4b3\") " pod="openstack/keystone-bootstrap-zbvhf" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.172927 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhvrq\" (UniqueName: \"kubernetes.io/projected/4bca3c7d-5b59-41d0-b441-546e4417718c-kube-api-access-rhvrq\") pod \"dnsmasq-dns-bbf5cc879-wrj6z\" (UID: \"4bca3c7d-5b59-41d0-b441-546e4417718c\") " pod="openstack/dnsmasq-dns-bbf5cc879-wrj6z" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.172955 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-combined-ca-bundle\") pod \"keystone-bootstrap-zbvhf\" (UID: \"e5d43f84-cbc2-43a7-877c-56d53b8de4b3\") " pod="openstack/keystone-bootstrap-zbvhf" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.172975 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4bca3c7d-5b59-41d0-b441-546e4417718c-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-wrj6z\" (UID: \"4bca3c7d-5b59-41d0-b441-546e4417718c\") " pod="openstack/dnsmasq-dns-bbf5cc879-wrj6z" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.172996 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4bca3c7d-5b59-41d0-b441-546e4417718c-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-wrj6z\" (UID: \"4bca3c7d-5b59-41d0-b441-546e4417718c\") " pod="openstack/dnsmasq-dns-bbf5cc879-wrj6z" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.173013 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4bca3c7d-5b59-41d0-b441-546e4417718c-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-wrj6z\" (UID: \"4bca3c7d-5b59-41d0-b441-546e4417718c\") " pod="openstack/dnsmasq-dns-bbf5cc879-wrj6z" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.173031 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bca3c7d-5b59-41d0-b441-546e4417718c-config\") pod \"dnsmasq-dns-bbf5cc879-wrj6z\" (UID: \"4bca3c7d-5b59-41d0-b441-546e4417718c\") " pod="openstack/dnsmasq-dns-bbf5cc879-wrj6z" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.173068 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-credential-keys\") pod \"keystone-bootstrap-zbvhf\" (UID: \"e5d43f84-cbc2-43a7-877c-56d53b8de4b3\") " pod="openstack/keystone-bootstrap-zbvhf" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.173083 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4bca3c7d-5b59-41d0-b441-546e4417718c-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-wrj6z\" (UID: \"4bca3c7d-5b59-41d0-b441-546e4417718c\") " pod="openstack/dnsmasq-dns-bbf5cc879-wrj6z" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.275078 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4bca3c7d-5b59-41d0-b441-546e4417718c-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-wrj6z\" (UID: \"4bca3c7d-5b59-41d0-b441-546e4417718c\") " pod="openstack/dnsmasq-dns-bbf5cc879-wrj6z" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.275132 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4bca3c7d-5b59-41d0-b441-546e4417718c-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-wrj6z\" (UID: \"4bca3c7d-5b59-41d0-b441-546e4417718c\") " pod="openstack/dnsmasq-dns-bbf5cc879-wrj6z" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.275173 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bca3c7d-5b59-41d0-b441-546e4417718c-config\") pod \"dnsmasq-dns-bbf5cc879-wrj6z\" (UID: \"4bca3c7d-5b59-41d0-b441-546e4417718c\") " pod="openstack/dnsmasq-dns-bbf5cc879-wrj6z" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.275217 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-credential-keys\") pod \"keystone-bootstrap-zbvhf\" (UID: \"e5d43f84-cbc2-43a7-877c-56d53b8de4b3\") " pod="openstack/keystone-bootstrap-zbvhf" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.275255 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4bca3c7d-5b59-41d0-b441-546e4417718c-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-wrj6z\" (UID: \"4bca3c7d-5b59-41d0-b441-546e4417718c\") " pod="openstack/dnsmasq-dns-bbf5cc879-wrj6z" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.275322 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsrk8\" (UniqueName: \"kubernetes.io/projected/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-kube-api-access-gsrk8\") pod \"keystone-bootstrap-zbvhf\" (UID: \"e5d43f84-cbc2-43a7-877c-56d53b8de4b3\") " pod="openstack/keystone-bootstrap-zbvhf" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.276209 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4bca3c7d-5b59-41d0-b441-546e4417718c-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-wrj6z\" (UID: \"4bca3c7d-5b59-41d0-b441-546e4417718c\") " pod="openstack/dnsmasq-dns-bbf5cc879-wrj6z" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.276321 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bca3c7d-5b59-41d0-b441-546e4417718c-config\") pod \"dnsmasq-dns-bbf5cc879-wrj6z\" (UID: \"4bca3c7d-5b59-41d0-b441-546e4417718c\") " pod="openstack/dnsmasq-dns-bbf5cc879-wrj6z" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.276339 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4bca3c7d-5b59-41d0-b441-546e4417718c-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-wrj6z\" (UID: \"4bca3c7d-5b59-41d0-b441-546e4417718c\") " pod="openstack/dnsmasq-dns-bbf5cc879-wrj6z" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.276410 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4bca3c7d-5b59-41d0-b441-546e4417718c-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-wrj6z\" (UID: \"4bca3c7d-5b59-41d0-b441-546e4417718c\") " pod="openstack/dnsmasq-dns-bbf5cc879-wrj6z" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.276457 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-config-data\") pod \"keystone-bootstrap-zbvhf\" (UID: \"e5d43f84-cbc2-43a7-877c-56d53b8de4b3\") " pod="openstack/keystone-bootstrap-zbvhf" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.276488 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-fernet-keys\") pod \"keystone-bootstrap-zbvhf\" (UID: \"e5d43f84-cbc2-43a7-877c-56d53b8de4b3\") " pod="openstack/keystone-bootstrap-zbvhf" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.276510 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-scripts\") pod \"keystone-bootstrap-zbvhf\" (UID: \"e5d43f84-cbc2-43a7-877c-56d53b8de4b3\") " pod="openstack/keystone-bootstrap-zbvhf" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.278281 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhvrq\" (UniqueName: \"kubernetes.io/projected/4bca3c7d-5b59-41d0-b441-546e4417718c-kube-api-access-rhvrq\") pod \"dnsmasq-dns-bbf5cc879-wrj6z\" (UID: \"4bca3c7d-5b59-41d0-b441-546e4417718c\") " pod="openstack/dnsmasq-dns-bbf5cc879-wrj6z" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.278323 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-combined-ca-bundle\") pod \"keystone-bootstrap-zbvhf\" (UID: \"e5d43f84-cbc2-43a7-877c-56d53b8de4b3\") " pod="openstack/keystone-bootstrap-zbvhf" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.278342 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4bca3c7d-5b59-41d0-b441-546e4417718c-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-wrj6z\" (UID: \"4bca3c7d-5b59-41d0-b441-546e4417718c\") " pod="openstack/dnsmasq-dns-bbf5cc879-wrj6z" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.278903 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4bca3c7d-5b59-41d0-b441-546e4417718c-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-wrj6z\" (UID: \"4bca3c7d-5b59-41d0-b441-546e4417718c\") " pod="openstack/dnsmasq-dns-bbf5cc879-wrj6z" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.279687 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-credential-keys\") pod \"keystone-bootstrap-zbvhf\" (UID: \"e5d43f84-cbc2-43a7-877c-56d53b8de4b3\") " pod="openstack/keystone-bootstrap-zbvhf" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.280534 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-fernet-keys\") pod \"keystone-bootstrap-zbvhf\" (UID: \"e5d43f84-cbc2-43a7-877c-56d53b8de4b3\") " pod="openstack/keystone-bootstrap-zbvhf" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.283471 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-combined-ca-bundle\") pod \"keystone-bootstrap-zbvhf\" (UID: \"e5d43f84-cbc2-43a7-877c-56d53b8de4b3\") " pod="openstack/keystone-bootstrap-zbvhf" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.289610 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsrk8\" (UniqueName: \"kubernetes.io/projected/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-kube-api-access-gsrk8\") pod \"keystone-bootstrap-zbvhf\" (UID: \"e5d43f84-cbc2-43a7-877c-56d53b8de4b3\") " pod="openstack/keystone-bootstrap-zbvhf" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.290488 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-scripts\") pod \"keystone-bootstrap-zbvhf\" (UID: \"e5d43f84-cbc2-43a7-877c-56d53b8de4b3\") " pod="openstack/keystone-bootstrap-zbvhf" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.293489 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhvrq\" (UniqueName: \"kubernetes.io/projected/4bca3c7d-5b59-41d0-b441-546e4417718c-kube-api-access-rhvrq\") pod \"dnsmasq-dns-bbf5cc879-wrj6z\" (UID: \"4bca3c7d-5b59-41d0-b441-546e4417718c\") " pod="openstack/dnsmasq-dns-bbf5cc879-wrj6z" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.299636 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-config-data\") pod \"keystone-bootstrap-zbvhf\" (UID: \"e5d43f84-cbc2-43a7-877c-56d53b8de4b3\") " pod="openstack/keystone-bootstrap-zbvhf" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.426581 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-wrj6z" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.440139 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zbvhf" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.850331 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6b588dd7f7-bbddl"] Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.869649 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-5zx65"] Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.870726 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-5zx65" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.871430 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6b588dd7f7-bbddl" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.905316 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.905424 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.905589 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-qmg7c" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.905850 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.906102 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-blc4m" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.906745 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.915713 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-5zx65"] Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.926839 4813 generic.go:334] "Generic (PLEG): container finished" podID="d1f29fa9-6e22-45be-97c4-4317914c1aa6" containerID="6e91d9bd84b543fe730a650e2cb0ad54ef7b2740d8aff51308221066d025c3cc" exitCode=0 Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.926881 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-gtr8j" event={"ID":"d1f29fa9-6e22-45be-97c4-4317914c1aa6","Type":"ContainerDied","Data":"6e91d9bd84b543fe730a650e2cb0ad54ef7b2740d8aff51308221066d025c3cc"} Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.944487 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.946718 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6b588dd7f7-bbddl"] Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.963662 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.965948 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.989635 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 16:02:36 crc kubenswrapper[4813]: I1206 16:02:36.989804 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.007322 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72532d82-6482-41ac-a461-c792c134dee3-combined-ca-bundle\") pod \"neutron-db-sync-5zx65\" (UID: \"72532d82-6482-41ac-a461-c792c134dee3\") " pod="openstack/neutron-db-sync-5zx65" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.007408 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1-horizon-secret-key\") pod \"horizon-6b588dd7f7-bbddl\" (UID: \"ef70d3a8-1fae-4950-b8c8-e12b48b7dac1\") " pod="openstack/horizon-6b588dd7f7-bbddl" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.007434 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/72532d82-6482-41ac-a461-c792c134dee3-config\") pod \"neutron-db-sync-5zx65\" (UID: \"72532d82-6482-41ac-a461-c792c134dee3\") " pod="openstack/neutron-db-sync-5zx65" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.007456 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6gk8\" (UniqueName: \"kubernetes.io/projected/72532d82-6482-41ac-a461-c792c134dee3-kube-api-access-l6gk8\") pod \"neutron-db-sync-5zx65\" (UID: \"72532d82-6482-41ac-a461-c792c134dee3\") " pod="openstack/neutron-db-sync-5zx65" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.007472 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1-logs\") pod \"horizon-6b588dd7f7-bbddl\" (UID: \"ef70d3a8-1fae-4950-b8c8-e12b48b7dac1\") " pod="openstack/horizon-6b588dd7f7-bbddl" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.007513 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cn8f\" (UniqueName: \"kubernetes.io/projected/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1-kube-api-access-2cn8f\") pod \"horizon-6b588dd7f7-bbddl\" (UID: \"ef70d3a8-1fae-4950-b8c8-e12b48b7dac1\") " pod="openstack/horizon-6b588dd7f7-bbddl" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.007529 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1-scripts\") pod \"horizon-6b588dd7f7-bbddl\" (UID: \"ef70d3a8-1fae-4950-b8c8-e12b48b7dac1\") " pod="openstack/horizon-6b588dd7f7-bbddl" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.007554 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1-config-data\") pod \"horizon-6b588dd7f7-bbddl\" (UID: \"ef70d3a8-1fae-4950-b8c8-e12b48b7dac1\") " pod="openstack/horizon-6b588dd7f7-bbddl" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.028135 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.108429 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72532d82-6482-41ac-a461-c792c134dee3-combined-ca-bundle\") pod \"neutron-db-sync-5zx65\" (UID: \"72532d82-6482-41ac-a461-c792c134dee3\") " pod="openstack/neutron-db-sync-5zx65" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.108490 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f539e8ab-300c-426b-98e7-f9e87a6787b5-run-httpd\") pod \"ceilometer-0\" (UID: \"f539e8ab-300c-426b-98e7-f9e87a6787b5\") " pod="openstack/ceilometer-0" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.108523 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f539e8ab-300c-426b-98e7-f9e87a6787b5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f539e8ab-300c-426b-98e7-f9e87a6787b5\") " pod="openstack/ceilometer-0" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.108542 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f539e8ab-300c-426b-98e7-f9e87a6787b5-scripts\") pod \"ceilometer-0\" (UID: \"f539e8ab-300c-426b-98e7-f9e87a6787b5\") " pod="openstack/ceilometer-0" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.108574 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1-horizon-secret-key\") pod \"horizon-6b588dd7f7-bbddl\" (UID: \"ef70d3a8-1fae-4950-b8c8-e12b48b7dac1\") " pod="openstack/horizon-6b588dd7f7-bbddl" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.108593 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f539e8ab-300c-426b-98e7-f9e87a6787b5-config-data\") pod \"ceilometer-0\" (UID: \"f539e8ab-300c-426b-98e7-f9e87a6787b5\") " pod="openstack/ceilometer-0" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.108609 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/72532d82-6482-41ac-a461-c792c134dee3-config\") pod \"neutron-db-sync-5zx65\" (UID: \"72532d82-6482-41ac-a461-c792c134dee3\") " pod="openstack/neutron-db-sync-5zx65" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.108627 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f539e8ab-300c-426b-98e7-f9e87a6787b5-log-httpd\") pod \"ceilometer-0\" (UID: \"f539e8ab-300c-426b-98e7-f9e87a6787b5\") " pod="openstack/ceilometer-0" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.108646 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1-logs\") pod \"horizon-6b588dd7f7-bbddl\" (UID: \"ef70d3a8-1fae-4950-b8c8-e12b48b7dac1\") " pod="openstack/horizon-6b588dd7f7-bbddl" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.108670 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6gk8\" (UniqueName: \"kubernetes.io/projected/72532d82-6482-41ac-a461-c792c134dee3-kube-api-access-l6gk8\") pod \"neutron-db-sync-5zx65\" (UID: \"72532d82-6482-41ac-a461-c792c134dee3\") " pod="openstack/neutron-db-sync-5zx65" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.108709 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f539e8ab-300c-426b-98e7-f9e87a6787b5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f539e8ab-300c-426b-98e7-f9e87a6787b5\") " pod="openstack/ceilometer-0" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.108728 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cn8f\" (UniqueName: \"kubernetes.io/projected/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1-kube-api-access-2cn8f\") pod \"horizon-6b588dd7f7-bbddl\" (UID: \"ef70d3a8-1fae-4950-b8c8-e12b48b7dac1\") " pod="openstack/horizon-6b588dd7f7-bbddl" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.108745 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1-scripts\") pod \"horizon-6b588dd7f7-bbddl\" (UID: \"ef70d3a8-1fae-4950-b8c8-e12b48b7dac1\") " pod="openstack/horizon-6b588dd7f7-bbddl" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.108761 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6tkg\" (UniqueName: \"kubernetes.io/projected/f539e8ab-300c-426b-98e7-f9e87a6787b5-kube-api-access-k6tkg\") pod \"ceilometer-0\" (UID: \"f539e8ab-300c-426b-98e7-f9e87a6787b5\") " pod="openstack/ceilometer-0" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.108784 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1-config-data\") pod \"horizon-6b588dd7f7-bbddl\" (UID: \"ef70d3a8-1fae-4950-b8c8-e12b48b7dac1\") " pod="openstack/horizon-6b588dd7f7-bbddl" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.109990 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1-config-data\") pod \"horizon-6b588dd7f7-bbddl\" (UID: \"ef70d3a8-1fae-4950-b8c8-e12b48b7dac1\") " pod="openstack/horizon-6b588dd7f7-bbddl" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.117474 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-qk8sl"] Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.118956 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1-horizon-secret-key\") pod \"horizon-6b588dd7f7-bbddl\" (UID: \"ef70d3a8-1fae-4950-b8c8-e12b48b7dac1\") " pod="openstack/horizon-6b588dd7f7-bbddl" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.119891 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72532d82-6482-41ac-a461-c792c134dee3-combined-ca-bundle\") pod \"neutron-db-sync-5zx65\" (UID: \"72532d82-6482-41ac-a461-c792c134dee3\") " pod="openstack/neutron-db-sync-5zx65" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.120132 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1-logs\") pod \"horizon-6b588dd7f7-bbddl\" (UID: \"ef70d3a8-1fae-4950-b8c8-e12b48b7dac1\") " pod="openstack/horizon-6b588dd7f7-bbddl" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.120878 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1-scripts\") pod \"horizon-6b588dd7f7-bbddl\" (UID: \"ef70d3a8-1fae-4950-b8c8-e12b48b7dac1\") " pod="openstack/horizon-6b588dd7f7-bbddl" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.124683 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-qk8sl" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.126974 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/72532d82-6482-41ac-a461-c792c134dee3-config\") pod \"neutron-db-sync-5zx65\" (UID: \"72532d82-6482-41ac-a461-c792c134dee3\") " pod="openstack/neutron-db-sync-5zx65" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.143616 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.143868 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-swv95" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.150117 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6gk8\" (UniqueName: \"kubernetes.io/projected/72532d82-6482-41ac-a461-c792c134dee3-kube-api-access-l6gk8\") pod \"neutron-db-sync-5zx65\" (UID: \"72532d82-6482-41ac-a461-c792c134dee3\") " pod="openstack/neutron-db-sync-5zx65" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.155037 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-qk8sl"] Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.188822 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cn8f\" (UniqueName: \"kubernetes.io/projected/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1-kube-api-access-2cn8f\") pod \"horizon-6b588dd7f7-bbddl\" (UID: \"ef70d3a8-1fae-4950-b8c8-e12b48b7dac1\") " pod="openstack/horizon-6b588dd7f7-bbddl" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.212633 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f539e8ab-300c-426b-98e7-f9e87a6787b5-run-httpd\") pod \"ceilometer-0\" (UID: \"f539e8ab-300c-426b-98e7-f9e87a6787b5\") " pod="openstack/ceilometer-0" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.212701 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f539e8ab-300c-426b-98e7-f9e87a6787b5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f539e8ab-300c-426b-98e7-f9e87a6787b5\") " pod="openstack/ceilometer-0" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.212734 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f539e8ab-300c-426b-98e7-f9e87a6787b5-scripts\") pod \"ceilometer-0\" (UID: \"f539e8ab-300c-426b-98e7-f9e87a6787b5\") " pod="openstack/ceilometer-0" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.212775 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5c1f8568-2368-445f-a2cb-01ab3a6d1564-db-sync-config-data\") pod \"barbican-db-sync-qk8sl\" (UID: \"5c1f8568-2368-445f-a2cb-01ab3a6d1564\") " pod="openstack/barbican-db-sync-qk8sl" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.212816 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f539e8ab-300c-426b-98e7-f9e87a6787b5-config-data\") pod \"ceilometer-0\" (UID: \"f539e8ab-300c-426b-98e7-f9e87a6787b5\") " pod="openstack/ceilometer-0" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.212844 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f539e8ab-300c-426b-98e7-f9e87a6787b5-log-httpd\") pod \"ceilometer-0\" (UID: \"f539e8ab-300c-426b-98e7-f9e87a6787b5\") " pod="openstack/ceilometer-0" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.212887 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkchj\" (UniqueName: \"kubernetes.io/projected/5c1f8568-2368-445f-a2cb-01ab3a6d1564-kube-api-access-jkchj\") pod \"barbican-db-sync-qk8sl\" (UID: \"5c1f8568-2368-445f-a2cb-01ab3a6d1564\") " pod="openstack/barbican-db-sync-qk8sl" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.212915 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c1f8568-2368-445f-a2cb-01ab3a6d1564-combined-ca-bundle\") pod \"barbican-db-sync-qk8sl\" (UID: \"5c1f8568-2368-445f-a2cb-01ab3a6d1564\") " pod="openstack/barbican-db-sync-qk8sl" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.212947 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f539e8ab-300c-426b-98e7-f9e87a6787b5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f539e8ab-300c-426b-98e7-f9e87a6787b5\") " pod="openstack/ceilometer-0" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.212977 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6tkg\" (UniqueName: \"kubernetes.io/projected/f539e8ab-300c-426b-98e7-f9e87a6787b5-kube-api-access-k6tkg\") pod \"ceilometer-0\" (UID: \"f539e8ab-300c-426b-98e7-f9e87a6787b5\") " pod="openstack/ceilometer-0" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.213177 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f539e8ab-300c-426b-98e7-f9e87a6787b5-run-httpd\") pod \"ceilometer-0\" (UID: \"f539e8ab-300c-426b-98e7-f9e87a6787b5\") " pod="openstack/ceilometer-0" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.213719 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f539e8ab-300c-426b-98e7-f9e87a6787b5-log-httpd\") pod \"ceilometer-0\" (UID: \"f539e8ab-300c-426b-98e7-f9e87a6787b5\") " pod="openstack/ceilometer-0" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.221549 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f539e8ab-300c-426b-98e7-f9e87a6787b5-config-data\") pod \"ceilometer-0\" (UID: \"f539e8ab-300c-426b-98e7-f9e87a6787b5\") " pod="openstack/ceilometer-0" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.236477 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f539e8ab-300c-426b-98e7-f9e87a6787b5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f539e8ab-300c-426b-98e7-f9e87a6787b5\") " pod="openstack/ceilometer-0" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.248963 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f539e8ab-300c-426b-98e7-f9e87a6787b5-scripts\") pod \"ceilometer-0\" (UID: \"f539e8ab-300c-426b-98e7-f9e87a6787b5\") " pod="openstack/ceilometer-0" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.254095 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f539e8ab-300c-426b-98e7-f9e87a6787b5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f539e8ab-300c-426b-98e7-f9e87a6787b5\") " pod="openstack/ceilometer-0" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.283529 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-5zx65" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.292078 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6b588dd7f7-bbddl" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.314102 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5c1f8568-2368-445f-a2cb-01ab3a6d1564-db-sync-config-data\") pod \"barbican-db-sync-qk8sl\" (UID: \"5c1f8568-2368-445f-a2cb-01ab3a6d1564\") " pod="openstack/barbican-db-sync-qk8sl" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.314390 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkchj\" (UniqueName: \"kubernetes.io/projected/5c1f8568-2368-445f-a2cb-01ab3a6d1564-kube-api-access-jkchj\") pod \"barbican-db-sync-qk8sl\" (UID: \"5c1f8568-2368-445f-a2cb-01ab3a6d1564\") " pod="openstack/barbican-db-sync-qk8sl" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.314485 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c1f8568-2368-445f-a2cb-01ab3a6d1564-combined-ca-bundle\") pod \"barbican-db-sync-qk8sl\" (UID: \"5c1f8568-2368-445f-a2cb-01ab3a6d1564\") " pod="openstack/barbican-db-sync-qk8sl" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.314824 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6tkg\" (UniqueName: \"kubernetes.io/projected/f539e8ab-300c-426b-98e7-f9e87a6787b5-kube-api-access-k6tkg\") pod \"ceilometer-0\" (UID: \"f539e8ab-300c-426b-98e7-f9e87a6787b5\") " pod="openstack/ceilometer-0" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.323941 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c1f8568-2368-445f-a2cb-01ab3a6d1564-combined-ca-bundle\") pod \"barbican-db-sync-qk8sl\" (UID: \"5c1f8568-2368-445f-a2cb-01ab3a6d1564\") " pod="openstack/barbican-db-sync-qk8sl" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.353393 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-lvw7n"] Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.360772 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-lvw7n" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.363743 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5c1f8568-2368-445f-a2cb-01ab3a6d1564-db-sync-config-data\") pod \"barbican-db-sync-qk8sl\" (UID: \"5c1f8568-2368-445f-a2cb-01ab3a6d1564\") " pod="openstack/barbican-db-sync-qk8sl" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.382012 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-wrj6z"] Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.395120 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.395312 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.413623 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkchj\" (UniqueName: \"kubernetes.io/projected/5c1f8568-2368-445f-a2cb-01ab3a6d1564-kube-api-access-jkchj\") pod \"barbican-db-sync-qk8sl\" (UID: \"5c1f8568-2368-445f-a2cb-01ab3a6d1564\") " pod="openstack/barbican-db-sync-qk8sl" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.416400 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-lvw7n"] Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.425820 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-wx499" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.418012 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-config-data\") pod \"cinder-db-sync-lvw7n\" (UID: \"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32\") " pod="openstack/cinder-db-sync-lvw7n" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.430747 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpz7c\" (UniqueName: \"kubernetes.io/projected/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-kube-api-access-cpz7c\") pod \"cinder-db-sync-lvw7n\" (UID: \"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32\") " pod="openstack/cinder-db-sync-lvw7n" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.430807 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-combined-ca-bundle\") pod \"cinder-db-sync-lvw7n\" (UID: \"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32\") " pod="openstack/cinder-db-sync-lvw7n" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.430968 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-etc-machine-id\") pod \"cinder-db-sync-lvw7n\" (UID: \"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32\") " pod="openstack/cinder-db-sync-lvw7n" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.431009 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-db-sync-config-data\") pod \"cinder-db-sync-lvw7n\" (UID: \"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32\") " pod="openstack/cinder-db-sync-lvw7n" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.431063 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-scripts\") pod \"cinder-db-sync-lvw7n\" (UID: \"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32\") " pod="openstack/cinder-db-sync-lvw7n" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.436675 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7f949586b5-5pzsg"] Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.438198 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f949586b5-5pzsg" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.452842 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-qk8sl" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.509976 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f949586b5-5pzsg"] Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.541134 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b697e541-1bb3-4619-8cc6-ce548717c741-horizon-secret-key\") pod \"horizon-7f949586b5-5pzsg\" (UID: \"b697e541-1bb3-4619-8cc6-ce548717c741\") " pod="openstack/horizon-7f949586b5-5pzsg" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.541196 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b697e541-1bb3-4619-8cc6-ce548717c741-config-data\") pod \"horizon-7f949586b5-5pzsg\" (UID: \"b697e541-1bb3-4619-8cc6-ce548717c741\") " pod="openstack/horizon-7f949586b5-5pzsg" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.541217 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpz7c\" (UniqueName: \"kubernetes.io/projected/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-kube-api-access-cpz7c\") pod \"cinder-db-sync-lvw7n\" (UID: \"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32\") " pod="openstack/cinder-db-sync-lvw7n" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.541241 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-combined-ca-bundle\") pod \"cinder-db-sync-lvw7n\" (UID: \"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32\") " pod="openstack/cinder-db-sync-lvw7n" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.541332 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95zr6\" (UniqueName: \"kubernetes.io/projected/b697e541-1bb3-4619-8cc6-ce548717c741-kube-api-access-95zr6\") pod \"horizon-7f949586b5-5pzsg\" (UID: \"b697e541-1bb3-4619-8cc6-ce548717c741\") " pod="openstack/horizon-7f949586b5-5pzsg" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.541352 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-etc-machine-id\") pod \"cinder-db-sync-lvw7n\" (UID: \"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32\") " pod="openstack/cinder-db-sync-lvw7n" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.541369 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b697e541-1bb3-4619-8cc6-ce548717c741-logs\") pod \"horizon-7f949586b5-5pzsg\" (UID: \"b697e541-1bb3-4619-8cc6-ce548717c741\") " pod="openstack/horizon-7f949586b5-5pzsg" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.541387 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-db-sync-config-data\") pod \"cinder-db-sync-lvw7n\" (UID: \"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32\") " pod="openstack/cinder-db-sync-lvw7n" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.541415 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-scripts\") pod \"cinder-db-sync-lvw7n\" (UID: \"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32\") " pod="openstack/cinder-db-sync-lvw7n" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.541446 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b697e541-1bb3-4619-8cc6-ce548717c741-scripts\") pod \"horizon-7f949586b5-5pzsg\" (UID: \"b697e541-1bb3-4619-8cc6-ce548717c741\") " pod="openstack/horizon-7f949586b5-5pzsg" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.541474 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-config-data\") pod \"cinder-db-sync-lvw7n\" (UID: \"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32\") " pod="openstack/cinder-db-sync-lvw7n" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.543789 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-etc-machine-id\") pod \"cinder-db-sync-lvw7n\" (UID: \"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32\") " pod="openstack/cinder-db-sync-lvw7n" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.548409 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-nrdxf"] Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.549686 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nrdxf" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.551382 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-combined-ca-bundle\") pod \"cinder-db-sync-lvw7n\" (UID: \"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32\") " pod="openstack/cinder-db-sync-lvw7n" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.557101 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-kvctl" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.557398 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.557582 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.559331 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-config-data\") pod \"cinder-db-sync-lvw7n\" (UID: \"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32\") " pod="openstack/cinder-db-sync-lvw7n" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.560374 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-db-sync-config-data\") pod \"cinder-db-sync-lvw7n\" (UID: \"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32\") " pod="openstack/cinder-db-sync-lvw7n" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.560716 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-scripts\") pod \"cinder-db-sync-lvw7n\" (UID: \"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32\") " pod="openstack/cinder-db-sync-lvw7n" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.619787 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.636744 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-nrdxf"] Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.639854 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpz7c\" (UniqueName: \"kubernetes.io/projected/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-kube-api-access-cpz7c\") pod \"cinder-db-sync-lvw7n\" (UID: \"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32\") " pod="openstack/cinder-db-sync-lvw7n" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.645353 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b697e541-1bb3-4619-8cc6-ce548717c741-config-data\") pod \"horizon-7f949586b5-5pzsg\" (UID: \"b697e541-1bb3-4619-8cc6-ce548717c741\") " pod="openstack/horizon-7f949586b5-5pzsg" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.645463 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0179e5b2-8549-4810-83ab-0acbff2a5c6c-combined-ca-bundle\") pod \"placement-db-sync-nrdxf\" (UID: \"0179e5b2-8549-4810-83ab-0acbff2a5c6c\") " pod="openstack/placement-db-sync-nrdxf" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.645560 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0179e5b2-8549-4810-83ab-0acbff2a5c6c-config-data\") pod \"placement-db-sync-nrdxf\" (UID: \"0179e5b2-8549-4810-83ab-0acbff2a5c6c\") " pod="openstack/placement-db-sync-nrdxf" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.645669 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0179e5b2-8549-4810-83ab-0acbff2a5c6c-scripts\") pod \"placement-db-sync-nrdxf\" (UID: \"0179e5b2-8549-4810-83ab-0acbff2a5c6c\") " pod="openstack/placement-db-sync-nrdxf" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.645754 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzm7x\" (UniqueName: \"kubernetes.io/projected/0179e5b2-8549-4810-83ab-0acbff2a5c6c-kube-api-access-pzm7x\") pod \"placement-db-sync-nrdxf\" (UID: \"0179e5b2-8549-4810-83ab-0acbff2a5c6c\") " pod="openstack/placement-db-sync-nrdxf" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.645836 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95zr6\" (UniqueName: \"kubernetes.io/projected/b697e541-1bb3-4619-8cc6-ce548717c741-kube-api-access-95zr6\") pod \"horizon-7f949586b5-5pzsg\" (UID: \"b697e541-1bb3-4619-8cc6-ce548717c741\") " pod="openstack/horizon-7f949586b5-5pzsg" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.645913 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b697e541-1bb3-4619-8cc6-ce548717c741-logs\") pod \"horizon-7f949586b5-5pzsg\" (UID: \"b697e541-1bb3-4619-8cc6-ce548717c741\") " pod="openstack/horizon-7f949586b5-5pzsg" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.646028 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b697e541-1bb3-4619-8cc6-ce548717c741-scripts\") pod \"horizon-7f949586b5-5pzsg\" (UID: \"b697e541-1bb3-4619-8cc6-ce548717c741\") " pod="openstack/horizon-7f949586b5-5pzsg" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.646118 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0179e5b2-8549-4810-83ab-0acbff2a5c6c-logs\") pod \"placement-db-sync-nrdxf\" (UID: \"0179e5b2-8549-4810-83ab-0acbff2a5c6c\") " pod="openstack/placement-db-sync-nrdxf" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.646186 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b697e541-1bb3-4619-8cc6-ce548717c741-horizon-secret-key\") pod \"horizon-7f949586b5-5pzsg\" (UID: \"b697e541-1bb3-4619-8cc6-ce548717c741\") " pod="openstack/horizon-7f949586b5-5pzsg" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.648702 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b697e541-1bb3-4619-8cc6-ce548717c741-config-data\") pod \"horizon-7f949586b5-5pzsg\" (UID: \"b697e541-1bb3-4619-8cc6-ce548717c741\") " pod="openstack/horizon-7f949586b5-5pzsg" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.649238 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b697e541-1bb3-4619-8cc6-ce548717c741-logs\") pod \"horizon-7f949586b5-5pzsg\" (UID: \"b697e541-1bb3-4619-8cc6-ce548717c741\") " pod="openstack/horizon-7f949586b5-5pzsg" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.649748 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b697e541-1bb3-4619-8cc6-ce548717c741-scripts\") pod \"horizon-7f949586b5-5pzsg\" (UID: \"b697e541-1bb3-4619-8cc6-ce548717c741\") " pod="openstack/horizon-7f949586b5-5pzsg" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.658019 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b697e541-1bb3-4619-8cc6-ce548717c741-horizon-secret-key\") pod \"horizon-7f949586b5-5pzsg\" (UID: \"b697e541-1bb3-4619-8cc6-ce548717c741\") " pod="openstack/horizon-7f949586b5-5pzsg" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.703801 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95zr6\" (UniqueName: \"kubernetes.io/projected/b697e541-1bb3-4619-8cc6-ce548717c741-kube-api-access-95zr6\") pod \"horizon-7f949586b5-5pzsg\" (UID: \"b697e541-1bb3-4619-8cc6-ce548717c741\") " pod="openstack/horizon-7f949586b5-5pzsg" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.712492 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-wrj6z"] Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.747435 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0179e5b2-8549-4810-83ab-0acbff2a5c6c-logs\") pod \"placement-db-sync-nrdxf\" (UID: \"0179e5b2-8549-4810-83ab-0acbff2a5c6c\") " pod="openstack/placement-db-sync-nrdxf" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.747488 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0179e5b2-8549-4810-83ab-0acbff2a5c6c-combined-ca-bundle\") pod \"placement-db-sync-nrdxf\" (UID: \"0179e5b2-8549-4810-83ab-0acbff2a5c6c\") " pod="openstack/placement-db-sync-nrdxf" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.747516 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0179e5b2-8549-4810-83ab-0acbff2a5c6c-config-data\") pod \"placement-db-sync-nrdxf\" (UID: \"0179e5b2-8549-4810-83ab-0acbff2a5c6c\") " pod="openstack/placement-db-sync-nrdxf" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.747546 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0179e5b2-8549-4810-83ab-0acbff2a5c6c-scripts\") pod \"placement-db-sync-nrdxf\" (UID: \"0179e5b2-8549-4810-83ab-0acbff2a5c6c\") " pod="openstack/placement-db-sync-nrdxf" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.747583 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzm7x\" (UniqueName: \"kubernetes.io/projected/0179e5b2-8549-4810-83ab-0acbff2a5c6c-kube-api-access-pzm7x\") pod \"placement-db-sync-nrdxf\" (UID: \"0179e5b2-8549-4810-83ab-0acbff2a5c6c\") " pod="openstack/placement-db-sync-nrdxf" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.748152 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0179e5b2-8549-4810-83ab-0acbff2a5c6c-logs\") pod \"placement-db-sync-nrdxf\" (UID: \"0179e5b2-8549-4810-83ab-0acbff2a5c6c\") " pod="openstack/placement-db-sync-nrdxf" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.814161 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0179e5b2-8549-4810-83ab-0acbff2a5c6c-combined-ca-bundle\") pod \"placement-db-sync-nrdxf\" (UID: \"0179e5b2-8549-4810-83ab-0acbff2a5c6c\") " pod="openstack/placement-db-sync-nrdxf" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.830021 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-lvw7n" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.879336 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f949586b5-5pzsg" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.893137 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzm7x\" (UniqueName: \"kubernetes.io/projected/0179e5b2-8549-4810-83ab-0acbff2a5c6c-kube-api-access-pzm7x\") pod \"placement-db-sync-nrdxf\" (UID: \"0179e5b2-8549-4810-83ab-0acbff2a5c6c\") " pod="openstack/placement-db-sync-nrdxf" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.893408 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.907116 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.917324 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0179e5b2-8549-4810-83ab-0acbff2a5c6c-scripts\") pod \"placement-db-sync-nrdxf\" (UID: \"0179e5b2-8549-4810-83ab-0acbff2a5c6c\") " pod="openstack/placement-db-sync-nrdxf" Dec 06 16:02:37 crc kubenswrapper[4813]: I1206 16:02:37.921005 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:37.953692 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0179e5b2-8549-4810-83ab-0acbff2a5c6c-config-data\") pod \"placement-db-sync-nrdxf\" (UID: \"0179e5b2-8549-4810-83ab-0acbff2a5c6c\") " pod="openstack/placement-db-sync-nrdxf" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:37.960957 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-gtr8j" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:37.935283 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:37.935329 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-sxdvk" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:37.965252 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.029841 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.065712 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1f29fa9-6e22-45be-97c4-4317914c1aa6-config\") pod \"d1f29fa9-6e22-45be-97c4-4317914c1aa6\" (UID: \"d1f29fa9-6e22-45be-97c4-4317914c1aa6\") " Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.065783 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1f29fa9-6e22-45be-97c4-4317914c1aa6-dns-svc\") pod \"d1f29fa9-6e22-45be-97c4-4317914c1aa6\" (UID: \"d1f29fa9-6e22-45be-97c4-4317914c1aa6\") " Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.065824 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d1f29fa9-6e22-45be-97c4-4317914c1aa6-ovsdbserver-nb\") pod \"d1f29fa9-6e22-45be-97c4-4317914c1aa6\" (UID: \"d1f29fa9-6e22-45be-97c4-4317914c1aa6\") " Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.065892 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htp6w\" (UniqueName: \"kubernetes.io/projected/d1f29fa9-6e22-45be-97c4-4317914c1aa6-kube-api-access-htp6w\") pod \"d1f29fa9-6e22-45be-97c4-4317914c1aa6\" (UID: \"d1f29fa9-6e22-45be-97c4-4317914c1aa6\") " Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.065945 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d1f29fa9-6e22-45be-97c4-4317914c1aa6-dns-swift-storage-0\") pod \"d1f29fa9-6e22-45be-97c4-4317914c1aa6\" (UID: \"d1f29fa9-6e22-45be-97c4-4317914c1aa6\") " Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.065963 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d1f29fa9-6e22-45be-97c4-4317914c1aa6-ovsdbserver-sb\") pod \"d1f29fa9-6e22-45be-97c4-4317914c1aa6\" (UID: \"d1f29fa9-6e22-45be-97c4-4317914c1aa6\") " Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.066142 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/372958b7-ef0c-448e-8acc-57a1b27a985f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " pod="openstack/glance-default-external-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.066206 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlz8h\" (UniqueName: \"kubernetes.io/projected/372958b7-ef0c-448e-8acc-57a1b27a985f-kube-api-access-rlz8h\") pod \"glance-default-external-api-0\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " pod="openstack/glance-default-external-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.066235 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/372958b7-ef0c-448e-8acc-57a1b27a985f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " pod="openstack/glance-default-external-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.066293 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/372958b7-ef0c-448e-8acc-57a1b27a985f-scripts\") pod \"glance-default-external-api-0\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " pod="openstack/glance-default-external-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.066309 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/372958b7-ef0c-448e-8acc-57a1b27a985f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " pod="openstack/glance-default-external-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.066325 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " pod="openstack/glance-default-external-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.066359 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/372958b7-ef0c-448e-8acc-57a1b27a985f-config-data\") pod \"glance-default-external-api-0\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " pod="openstack/glance-default-external-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.066406 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/372958b7-ef0c-448e-8acc-57a1b27a985f-logs\") pod \"glance-default-external-api-0\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " pod="openstack/glance-default-external-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.101179 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-rrfrw"] Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.101563 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1f29fa9-6e22-45be-97c4-4317914c1aa6-kube-api-access-htp6w" (OuterVolumeSpecName: "kube-api-access-htp6w") pod "d1f29fa9-6e22-45be-97c4-4317914c1aa6" (UID: "d1f29fa9-6e22-45be-97c4-4317914c1aa6"). InnerVolumeSpecName "kube-api-access-htp6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:02:38 crc kubenswrapper[4813]: E1206 16:02:38.101640 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1f29fa9-6e22-45be-97c4-4317914c1aa6" containerName="init" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.101654 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1f29fa9-6e22-45be-97c4-4317914c1aa6" containerName="init" Dec 06 16:02:38 crc kubenswrapper[4813]: E1206 16:02:38.101684 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1f29fa9-6e22-45be-97c4-4317914c1aa6" containerName="dnsmasq-dns" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.101690 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1f29fa9-6e22-45be-97c4-4317914c1aa6" containerName="dnsmasq-dns" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.101858 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1f29fa9-6e22-45be-97c4-4317914c1aa6" containerName="dnsmasq-dns" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.113017 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-rrfrw"] Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.113122 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.118354 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-wrj6z" event={"ID":"4bca3c7d-5b59-41d0-b441-546e4417718c","Type":"ContainerStarted","Data":"32281009b2f7102ab9abcdc721777b01f06f373141105f06b4f2210abcf2b072"} Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.146675 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-zbvhf"] Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.165775 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-gtr8j" event={"ID":"d1f29fa9-6e22-45be-97c4-4317914c1aa6","Type":"ContainerDied","Data":"8d45f528928bcc87089be5eb4ef976ce4f2a8ea362b0dacc88d08611e4cb3fbd"} Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.165880 4813 scope.go:117] "RemoveContainer" containerID="6e91d9bd84b543fe730a650e2cb0ad54ef7b2740d8aff51308221066d025c3cc" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.166006 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-gtr8j" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.167083 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/efca4ca3-8021-4417-83cc-1b26f13f97bd-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-rrfrw\" (UID: \"efca4ca3-8021-4417-83cc-1b26f13f97bd\") " pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.167116 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/372958b7-ef0c-448e-8acc-57a1b27a985f-config-data\") pod \"glance-default-external-api-0\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " pod="openstack/glance-default-external-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.167144 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/efca4ca3-8021-4417-83cc-1b26f13f97bd-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-rrfrw\" (UID: \"efca4ca3-8021-4417-83cc-1b26f13f97bd\") " pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.167185 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/372958b7-ef0c-448e-8acc-57a1b27a985f-logs\") pod \"glance-default-external-api-0\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " pod="openstack/glance-default-external-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.167211 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/372958b7-ef0c-448e-8acc-57a1b27a985f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " pod="openstack/glance-default-external-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.167227 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/efca4ca3-8021-4417-83cc-1b26f13f97bd-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-rrfrw\" (UID: \"efca4ca3-8021-4417-83cc-1b26f13f97bd\") " pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.167252 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efca4ca3-8021-4417-83cc-1b26f13f97bd-config\") pod \"dnsmasq-dns-56df8fb6b7-rrfrw\" (UID: \"efca4ca3-8021-4417-83cc-1b26f13f97bd\") " pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.167296 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlz8h\" (UniqueName: \"kubernetes.io/projected/372958b7-ef0c-448e-8acc-57a1b27a985f-kube-api-access-rlz8h\") pod \"glance-default-external-api-0\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " pod="openstack/glance-default-external-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.167320 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/372958b7-ef0c-448e-8acc-57a1b27a985f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " pod="openstack/glance-default-external-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.167350 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/efca4ca3-8021-4417-83cc-1b26f13f97bd-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-rrfrw\" (UID: \"efca4ca3-8021-4417-83cc-1b26f13f97bd\") " pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.167376 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/372958b7-ef0c-448e-8acc-57a1b27a985f-scripts\") pod \"glance-default-external-api-0\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " pod="openstack/glance-default-external-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.167391 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/372958b7-ef0c-448e-8acc-57a1b27a985f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " pod="openstack/glance-default-external-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.167406 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " pod="openstack/glance-default-external-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.167429 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxpln\" (UniqueName: \"kubernetes.io/projected/efca4ca3-8021-4417-83cc-1b26f13f97bd-kube-api-access-wxpln\") pod \"dnsmasq-dns-56df8fb6b7-rrfrw\" (UID: \"efca4ca3-8021-4417-83cc-1b26f13f97bd\") " pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.167470 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htp6w\" (UniqueName: \"kubernetes.io/projected/d1f29fa9-6e22-45be-97c4-4317914c1aa6-kube-api-access-htp6w\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.168397 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/372958b7-ef0c-448e-8acc-57a1b27a985f-logs\") pod \"glance-default-external-api-0\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " pod="openstack/glance-default-external-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.173124 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/372958b7-ef0c-448e-8acc-57a1b27a985f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " pod="openstack/glance-default-external-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.173387 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-external-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.207925 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.209007 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/372958b7-ef0c-448e-8acc-57a1b27a985f-config-data\") pod \"glance-default-external-api-0\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " pod="openstack/glance-default-external-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.211308 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/372958b7-ef0c-448e-8acc-57a1b27a985f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " pod="openstack/glance-default-external-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.221670 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.221698 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zbvhf" event={"ID":"e5d43f84-cbc2-43a7-877c-56d53b8de4b3","Type":"ContainerStarted","Data":"b48a4a245b0a4a48d62913b12ddba81cb7c75aea32427ae2dd8da1ec0df66aad"} Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.221788 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.230683 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/372958b7-ef0c-448e-8acc-57a1b27a985f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " pod="openstack/glance-default-external-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.231086 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.231251 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.231748 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/372958b7-ef0c-448e-8acc-57a1b27a985f-scripts\") pod \"glance-default-external-api-0\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " pod="openstack/glance-default-external-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.240813 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlz8h\" (UniqueName: \"kubernetes.io/projected/372958b7-ef0c-448e-8acc-57a1b27a985f-kube-api-access-rlz8h\") pod \"glance-default-external-api-0\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " pod="openstack/glance-default-external-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.267496 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nrdxf" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.273139 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/efca4ca3-8021-4417-83cc-1b26f13f97bd-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-rrfrw\" (UID: \"efca4ca3-8021-4417-83cc-1b26f13f97bd\") " pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.273205 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4340aaa-8da3-4293-87f2-d49f43ee01bd-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.274522 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/efca4ca3-8021-4417-83cc-1b26f13f97bd-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-rrfrw\" (UID: \"efca4ca3-8021-4417-83cc-1b26f13f97bd\") " pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.274583 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efca4ca3-8021-4417-83cc-1b26f13f97bd-config\") pod \"dnsmasq-dns-56df8fb6b7-rrfrw\" (UID: \"efca4ca3-8021-4417-83cc-1b26f13f97bd\") " pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.274607 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c4340aaa-8da3-4293-87f2-d49f43ee01bd-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.274640 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwq7q\" (UniqueName: \"kubernetes.io/projected/c4340aaa-8da3-4293-87f2-d49f43ee01bd-kube-api-access-vwq7q\") pod \"glance-default-internal-api-0\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.274660 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.274697 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4340aaa-8da3-4293-87f2-d49f43ee01bd-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.274745 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4340aaa-8da3-4293-87f2-d49f43ee01bd-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.274769 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/efca4ca3-8021-4417-83cc-1b26f13f97bd-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-rrfrw\" (UID: \"efca4ca3-8021-4417-83cc-1b26f13f97bd\") " pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.274830 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4340aaa-8da3-4293-87f2-d49f43ee01bd-logs\") pod \"glance-default-internal-api-0\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.274847 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4340aaa-8da3-4293-87f2-d49f43ee01bd-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.274882 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxpln\" (UniqueName: \"kubernetes.io/projected/efca4ca3-8021-4417-83cc-1b26f13f97bd-kube-api-access-wxpln\") pod \"dnsmasq-dns-56df8fb6b7-rrfrw\" (UID: \"efca4ca3-8021-4417-83cc-1b26f13f97bd\") " pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.274907 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/efca4ca3-8021-4417-83cc-1b26f13f97bd-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-rrfrw\" (UID: \"efca4ca3-8021-4417-83cc-1b26f13f97bd\") " pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.275239 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/efca4ca3-8021-4417-83cc-1b26f13f97bd-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-rrfrw\" (UID: \"efca4ca3-8021-4417-83cc-1b26f13f97bd\") " pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.276004 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/efca4ca3-8021-4417-83cc-1b26f13f97bd-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-rrfrw\" (UID: \"efca4ca3-8021-4417-83cc-1b26f13f97bd\") " pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.276523 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efca4ca3-8021-4417-83cc-1b26f13f97bd-config\") pod \"dnsmasq-dns-56df8fb6b7-rrfrw\" (UID: \"efca4ca3-8021-4417-83cc-1b26f13f97bd\") " pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.277014 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/efca4ca3-8021-4417-83cc-1b26f13f97bd-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-rrfrw\" (UID: \"efca4ca3-8021-4417-83cc-1b26f13f97bd\") " pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.324129 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/efca4ca3-8021-4417-83cc-1b26f13f97bd-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-rrfrw\" (UID: \"efca4ca3-8021-4417-83cc-1b26f13f97bd\") " pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.363441 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxpln\" (UniqueName: \"kubernetes.io/projected/efca4ca3-8021-4417-83cc-1b26f13f97bd-kube-api-access-wxpln\") pod \"dnsmasq-dns-56df8fb6b7-rrfrw\" (UID: \"efca4ca3-8021-4417-83cc-1b26f13f97bd\") " pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.379192 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4340aaa-8da3-4293-87f2-d49f43ee01bd-logs\") pod \"glance-default-internal-api-0\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.381176 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4340aaa-8da3-4293-87f2-d49f43ee01bd-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.381308 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4340aaa-8da3-4293-87f2-d49f43ee01bd-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.381041 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4340aaa-8da3-4293-87f2-d49f43ee01bd-logs\") pod \"glance-default-internal-api-0\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.383592 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c4340aaa-8da3-4293-87f2-d49f43ee01bd-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.383624 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwq7q\" (UniqueName: \"kubernetes.io/projected/c4340aaa-8da3-4293-87f2-d49f43ee01bd-kube-api-access-vwq7q\") pod \"glance-default-internal-api-0\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.383645 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.383698 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4340aaa-8da3-4293-87f2-d49f43ee01bd-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.383743 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4340aaa-8da3-4293-87f2-d49f43ee01bd-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.384380 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c4340aaa-8da3-4293-87f2-d49f43ee01bd-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.384682 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.397595 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4340aaa-8da3-4293-87f2-d49f43ee01bd-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.402959 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4340aaa-8da3-4293-87f2-d49f43ee01bd-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.421841 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4340aaa-8da3-4293-87f2-d49f43ee01bd-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.427830 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4340aaa-8da3-4293-87f2-d49f43ee01bd-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.436751 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwq7q\" (UniqueName: \"kubernetes.io/projected/c4340aaa-8da3-4293-87f2-d49f43ee01bd-kube-api-access-vwq7q\") pod \"glance-default-internal-api-0\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.481578 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1f29fa9-6e22-45be-97c4-4317914c1aa6-config" (OuterVolumeSpecName: "config") pod "d1f29fa9-6e22-45be-97c4-4317914c1aa6" (UID: "d1f29fa9-6e22-45be-97c4-4317914c1aa6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.487053 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1f29fa9-6e22-45be-97c4-4317914c1aa6-config\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.494314 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " pod="openstack/glance-default-external-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.545770 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1f29fa9-6e22-45be-97c4-4317914c1aa6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d1f29fa9-6e22-45be-97c4-4317914c1aa6" (UID: "d1f29fa9-6e22-45be-97c4-4317914c1aa6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.552968 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.553551 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1f29fa9-6e22-45be-97c4-4317914c1aa6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d1f29fa9-6e22-45be-97c4-4317914c1aa6" (UID: "d1f29fa9-6e22-45be-97c4-4317914c1aa6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.556322 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1f29fa9-6e22-45be-97c4-4317914c1aa6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d1f29fa9-6e22-45be-97c4-4317914c1aa6" (UID: "d1f29fa9-6e22-45be-97c4-4317914c1aa6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.586087 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1f29fa9-6e22-45be-97c4-4317914c1aa6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d1f29fa9-6e22-45be-97c4-4317914c1aa6" (UID: "d1f29fa9-6e22-45be-97c4-4317914c1aa6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.587889 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d1f29fa9-6e22-45be-97c4-4317914c1aa6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.587917 4813 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d1f29fa9-6e22-45be-97c4-4317914c1aa6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.587926 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d1f29fa9-6e22-45be-97c4-4317914c1aa6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.587936 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1f29fa9-6e22-45be-97c4-4317914c1aa6-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.629282 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.629868 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.639009 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-qk8sl"] Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.651173 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.669089 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6b588dd7f7-bbddl"] Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.710090 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-5zx65"] Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.715934 4813 scope.go:117] "RemoveContainer" containerID="95d42e55f4236eeb7fa3f110922fb389ad84293e533194f9070db39b74b0e7e9" Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.836378 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-gtr8j"] Dec 06 16:02:38 crc kubenswrapper[4813]: I1206 16:02:38.869733 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-gtr8j"] Dec 06 16:02:39 crc kubenswrapper[4813]: I1206 16:02:39.028360 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f949586b5-5pzsg"] Dec 06 16:02:39 crc kubenswrapper[4813]: I1206 16:02:39.044320 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:02:39 crc kubenswrapper[4813]: I1206 16:02:39.072406 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-lvw7n"] Dec 06 16:02:39 crc kubenswrapper[4813]: I1206 16:02:39.121125 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-nrdxf"] Dec 06 16:02:39 crc kubenswrapper[4813]: I1206 16:02:39.237417 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-qk8sl" event={"ID":"5c1f8568-2368-445f-a2cb-01ab3a6d1564","Type":"ContainerStarted","Data":"8d0dbd6457572bd8d3cecec5e9bc1befe436860af2eb9468111eb2dcaab364f9"} Dec 06 16:02:39 crc kubenswrapper[4813]: I1206 16:02:39.246828 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-lvw7n" event={"ID":"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32","Type":"ContainerStarted","Data":"22d0a046133d14bb5dc9621fe51f7f033ea8f91fedd722681ab075a659d6a00c"} Dec 06 16:02:39 crc kubenswrapper[4813]: I1206 16:02:39.250812 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f949586b5-5pzsg" event={"ID":"b697e541-1bb3-4619-8cc6-ce548717c741","Type":"ContainerStarted","Data":"b88c6e92ef8019f49b583d9c183ae296a31a09108e33269f9efa157e3fdeddb1"} Dec 06 16:02:39 crc kubenswrapper[4813]: I1206 16:02:39.264074 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zbvhf" event={"ID":"e5d43f84-cbc2-43a7-877c-56d53b8de4b3","Type":"ContainerStarted","Data":"bddf9da2a88bc9810142b3c40b3de2ff32912f6eb820812c93b8fa810c085281"} Dec 06 16:02:39 crc kubenswrapper[4813]: I1206 16:02:39.273108 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6b588dd7f7-bbddl" event={"ID":"ef70d3a8-1fae-4950-b8c8-e12b48b7dac1","Type":"ContainerStarted","Data":"d899bfdaf7bd3ace3f0c065abdf4067ae59ddcd2d40e92edc122c4a203fd91cc"} Dec 06 16:02:39 crc kubenswrapper[4813]: I1206 16:02:39.275376 4813 generic.go:334] "Generic (PLEG): container finished" podID="4bca3c7d-5b59-41d0-b441-546e4417718c" containerID="e0c8d9040f459a36e067dbe77c3cee3b62634d9ff45175e6874148147b24409e" exitCode=0 Dec 06 16:02:39 crc kubenswrapper[4813]: I1206 16:02:39.275424 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-wrj6z" event={"ID":"4bca3c7d-5b59-41d0-b441-546e4417718c","Type":"ContainerDied","Data":"e0c8d9040f459a36e067dbe77c3cee3b62634d9ff45175e6874148147b24409e"} Dec 06 16:02:39 crc kubenswrapper[4813]: I1206 16:02:39.305548 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-5zx65" event={"ID":"72532d82-6482-41ac-a461-c792c134dee3","Type":"ContainerStarted","Data":"c421e9528a587793ae5ac6d5f50537230cf1f38bb2b5734e64d67f19d34e4772"} Dec 06 16:02:39 crc kubenswrapper[4813]: I1206 16:02:39.305589 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-5zx65" event={"ID":"72532d82-6482-41ac-a461-c792c134dee3","Type":"ContainerStarted","Data":"4ba2406e5efc299c12ecae4c70e3ed7df3038b736e2d464e2c419bdc3392bcb2"} Dec 06 16:02:39 crc kubenswrapper[4813]: I1206 16:02:39.312501 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-zbvhf" podStartSLOduration=3.312482999 podStartE2EDuration="3.312482999s" podCreationTimestamp="2025-12-06 16:02:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:02:39.306483581 +0000 UTC m=+999.197363157" watchObservedRunningTime="2025-12-06 16:02:39.312482999 +0000 UTC m=+999.203362575" Dec 06 16:02:39 crc kubenswrapper[4813]: I1206 16:02:39.335675 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-5zx65" podStartSLOduration=3.335656463 podStartE2EDuration="3.335656463s" podCreationTimestamp="2025-12-06 16:02:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:02:39.333420174 +0000 UTC m=+999.224299750" watchObservedRunningTime="2025-12-06 16:02:39.335656463 +0000 UTC m=+999.226536039" Dec 06 16:02:39 crc kubenswrapper[4813]: I1206 16:02:39.350292 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nrdxf" event={"ID":"0179e5b2-8549-4810-83ab-0acbff2a5c6c","Type":"ContainerStarted","Data":"7f9d76ce538710b4003662dfeefe3596f9f64e19669a9c29903182fd9bfb2d9a"} Dec 06 16:02:39 crc kubenswrapper[4813]: I1206 16:02:39.367885 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f539e8ab-300c-426b-98e7-f9e87a6787b5","Type":"ContainerStarted","Data":"970e7c7a4af6f21c2152801c05d4c15dfb412dea047584ad9886f3dde225be9c"} Dec 06 16:02:39 crc kubenswrapper[4813]: I1206 16:02:39.409171 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 16:02:39 crc kubenswrapper[4813]: W1206 16:02:39.419829 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod372958b7_ef0c_448e_8acc_57a1b27a985f.slice/crio-2c70f8c53d8879b8a98ba274d20289687daf28521627dd9e5025946c0082699c WatchSource:0}: Error finding container 2c70f8c53d8879b8a98ba274d20289687daf28521627dd9e5025946c0082699c: Status 404 returned error can't find the container with id 2c70f8c53d8879b8a98ba274d20289687daf28521627dd9e5025946c0082699c Dec 06 16:02:39 crc kubenswrapper[4813]: I1206 16:02:39.531155 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-rrfrw"] Dec 06 16:02:39 crc kubenswrapper[4813]: I1206 16:02:39.797845 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 16:02:39 crc kubenswrapper[4813]: W1206 16:02:39.927060 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4340aaa_8da3_4293_87f2_d49f43ee01bd.slice/crio-4409c57b9e5fee22f6438ee07f0bd9b1da473524285dc380bfb6dd1c8c70defa WatchSource:0}: Error finding container 4409c57b9e5fee22f6438ee07f0bd9b1da473524285dc380bfb6dd1c8c70defa: Status 404 returned error can't find the container with id 4409c57b9e5fee22f6438ee07f0bd9b1da473524285dc380bfb6dd1c8c70defa Dec 06 16:02:39 crc kubenswrapper[4813]: I1206 16:02:39.966196 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-wrj6z" Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.034409 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bca3c7d-5b59-41d0-b441-546e4417718c-config\") pod \"4bca3c7d-5b59-41d0-b441-546e4417718c\" (UID: \"4bca3c7d-5b59-41d0-b441-546e4417718c\") " Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.034485 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhvrq\" (UniqueName: \"kubernetes.io/projected/4bca3c7d-5b59-41d0-b441-546e4417718c-kube-api-access-rhvrq\") pod \"4bca3c7d-5b59-41d0-b441-546e4417718c\" (UID: \"4bca3c7d-5b59-41d0-b441-546e4417718c\") " Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.034504 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4bca3c7d-5b59-41d0-b441-546e4417718c-ovsdbserver-nb\") pod \"4bca3c7d-5b59-41d0-b441-546e4417718c\" (UID: \"4bca3c7d-5b59-41d0-b441-546e4417718c\") " Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.034533 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4bca3c7d-5b59-41d0-b441-546e4417718c-dns-svc\") pod \"4bca3c7d-5b59-41d0-b441-546e4417718c\" (UID: \"4bca3c7d-5b59-41d0-b441-546e4417718c\") " Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.034549 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4bca3c7d-5b59-41d0-b441-546e4417718c-dns-swift-storage-0\") pod \"4bca3c7d-5b59-41d0-b441-546e4417718c\" (UID: \"4bca3c7d-5b59-41d0-b441-546e4417718c\") " Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.034591 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4bca3c7d-5b59-41d0-b441-546e4417718c-ovsdbserver-sb\") pod \"4bca3c7d-5b59-41d0-b441-546e4417718c\" (UID: \"4bca3c7d-5b59-41d0-b441-546e4417718c\") " Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.065176 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bca3c7d-5b59-41d0-b441-546e4417718c-kube-api-access-rhvrq" (OuterVolumeSpecName: "kube-api-access-rhvrq") pod "4bca3c7d-5b59-41d0-b441-546e4417718c" (UID: "4bca3c7d-5b59-41d0-b441-546e4417718c"). InnerVolumeSpecName "kube-api-access-rhvrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.089319 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bca3c7d-5b59-41d0-b441-546e4417718c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4bca3c7d-5b59-41d0-b441-546e4417718c" (UID: "4bca3c7d-5b59-41d0-b441-546e4417718c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.093969 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bca3c7d-5b59-41d0-b441-546e4417718c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4bca3c7d-5b59-41d0-b441-546e4417718c" (UID: "4bca3c7d-5b59-41d0-b441-546e4417718c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.099858 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bca3c7d-5b59-41d0-b441-546e4417718c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4bca3c7d-5b59-41d0-b441-546e4417718c" (UID: "4bca3c7d-5b59-41d0-b441-546e4417718c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.103600 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bca3c7d-5b59-41d0-b441-546e4417718c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4bca3c7d-5b59-41d0-b441-546e4417718c" (UID: "4bca3c7d-5b59-41d0-b441-546e4417718c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.110823 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bca3c7d-5b59-41d0-b441-546e4417718c-config" (OuterVolumeSpecName: "config") pod "4bca3c7d-5b59-41d0-b441-546e4417718c" (UID: "4bca3c7d-5b59-41d0-b441-546e4417718c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.145055 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bca3c7d-5b59-41d0-b441-546e4417718c-config\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.145083 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhvrq\" (UniqueName: \"kubernetes.io/projected/4bca3c7d-5b59-41d0-b441-546e4417718c-kube-api-access-rhvrq\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.145093 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4bca3c7d-5b59-41d0-b441-546e4417718c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.145101 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4bca3c7d-5b59-41d0-b441-546e4417718c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.145109 4813 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4bca3c7d-5b59-41d0-b441-546e4417718c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.145138 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4bca3c7d-5b59-41d0-b441-546e4417718c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.429317 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-wrj6z" event={"ID":"4bca3c7d-5b59-41d0-b441-546e4417718c","Type":"ContainerDied","Data":"32281009b2f7102ab9abcdc721777b01f06f373141105f06b4f2210abcf2b072"} Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.429662 4813 scope.go:117] "RemoveContainer" containerID="e0c8d9040f459a36e067dbe77c3cee3b62634d9ff45175e6874148147b24409e" Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.429780 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-wrj6z" Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.441397 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c4340aaa-8da3-4293-87f2-d49f43ee01bd","Type":"ContainerStarted","Data":"4409c57b9e5fee22f6438ee07f0bd9b1da473524285dc380bfb6dd1c8c70defa"} Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.483890 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"372958b7-ef0c-448e-8acc-57a1b27a985f","Type":"ContainerStarted","Data":"2c70f8c53d8879b8a98ba274d20289687daf28521627dd9e5025946c0082699c"} Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.539973 4813 generic.go:334] "Generic (PLEG): container finished" podID="efca4ca3-8021-4417-83cc-1b26f13f97bd" containerID="15ed9e2cd1cd900278014cb261390fa644c9ac32bac10da79943a454e74702cd" exitCode=0 Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.652545 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1f29fa9-6e22-45be-97c4-4317914c1aa6" path="/var/lib/kubelet/pods/d1f29fa9-6e22-45be-97c4-4317914c1aa6/volumes" Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.678689 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" event={"ID":"efca4ca3-8021-4417-83cc-1b26f13f97bd","Type":"ContainerDied","Data":"15ed9e2cd1cd900278014cb261390fa644c9ac32bac10da79943a454e74702cd"} Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.690943 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-wrj6z"] Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.691196 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" event={"ID":"efca4ca3-8021-4417-83cc-1b26f13f97bd","Type":"ContainerStarted","Data":"d4857fdf061b281100b9d4b234170eab196bc39a7280eb9072f484773c83bc11"} Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.691295 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.741416 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-wrj6z"] Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.786959 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6b588dd7f7-bbddl"] Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.811562 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.889298 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-86c95686f7-d9tjn"] Dec 06 16:02:40 crc kubenswrapper[4813]: E1206 16:02:40.889947 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bca3c7d-5b59-41d0-b441-546e4417718c" containerName="init" Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.890016 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bca3c7d-5b59-41d0-b441-546e4417718c" containerName="init" Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.890253 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bca3c7d-5b59-41d0-b441-546e4417718c" containerName="init" Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.891409 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-86c95686f7-d9tjn" Dec 06 16:02:40 crc kubenswrapper[4813]: I1206 16:02:40.932982 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-86c95686f7-d9tjn"] Dec 06 16:02:41 crc kubenswrapper[4813]: I1206 16:02:41.005782 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 16:02:41 crc kubenswrapper[4813]: I1206 16:02:41.014551 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4845d02f-a80d-4954-bede-716091ba57b8-config-data\") pod \"horizon-86c95686f7-d9tjn\" (UID: \"4845d02f-a80d-4954-bede-716091ba57b8\") " pod="openstack/horizon-86c95686f7-d9tjn" Dec 06 16:02:41 crc kubenswrapper[4813]: I1206 16:02:41.014622 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsqxj\" (UniqueName: \"kubernetes.io/projected/4845d02f-a80d-4954-bede-716091ba57b8-kube-api-access-bsqxj\") pod \"horizon-86c95686f7-d9tjn\" (UID: \"4845d02f-a80d-4954-bede-716091ba57b8\") " pod="openstack/horizon-86c95686f7-d9tjn" Dec 06 16:02:41 crc kubenswrapper[4813]: I1206 16:02:41.014681 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4845d02f-a80d-4954-bede-716091ba57b8-logs\") pod \"horizon-86c95686f7-d9tjn\" (UID: \"4845d02f-a80d-4954-bede-716091ba57b8\") " pod="openstack/horizon-86c95686f7-d9tjn" Dec 06 16:02:41 crc kubenswrapper[4813]: I1206 16:02:41.014775 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4845d02f-a80d-4954-bede-716091ba57b8-scripts\") pod \"horizon-86c95686f7-d9tjn\" (UID: \"4845d02f-a80d-4954-bede-716091ba57b8\") " pod="openstack/horizon-86c95686f7-d9tjn" Dec 06 16:02:41 crc kubenswrapper[4813]: I1206 16:02:41.014882 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4845d02f-a80d-4954-bede-716091ba57b8-horizon-secret-key\") pod \"horizon-86c95686f7-d9tjn\" (UID: \"4845d02f-a80d-4954-bede-716091ba57b8\") " pod="openstack/horizon-86c95686f7-d9tjn" Dec 06 16:02:41 crc kubenswrapper[4813]: I1206 16:02:41.116833 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4845d02f-a80d-4954-bede-716091ba57b8-config-data\") pod \"horizon-86c95686f7-d9tjn\" (UID: \"4845d02f-a80d-4954-bede-716091ba57b8\") " pod="openstack/horizon-86c95686f7-d9tjn" Dec 06 16:02:41 crc kubenswrapper[4813]: I1206 16:02:41.117102 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsqxj\" (UniqueName: \"kubernetes.io/projected/4845d02f-a80d-4954-bede-716091ba57b8-kube-api-access-bsqxj\") pod \"horizon-86c95686f7-d9tjn\" (UID: \"4845d02f-a80d-4954-bede-716091ba57b8\") " pod="openstack/horizon-86c95686f7-d9tjn" Dec 06 16:02:41 crc kubenswrapper[4813]: I1206 16:02:41.117133 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4845d02f-a80d-4954-bede-716091ba57b8-logs\") pod \"horizon-86c95686f7-d9tjn\" (UID: \"4845d02f-a80d-4954-bede-716091ba57b8\") " pod="openstack/horizon-86c95686f7-d9tjn" Dec 06 16:02:41 crc kubenswrapper[4813]: I1206 16:02:41.117174 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4845d02f-a80d-4954-bede-716091ba57b8-scripts\") pod \"horizon-86c95686f7-d9tjn\" (UID: \"4845d02f-a80d-4954-bede-716091ba57b8\") " pod="openstack/horizon-86c95686f7-d9tjn" Dec 06 16:02:41 crc kubenswrapper[4813]: I1206 16:02:41.117216 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4845d02f-a80d-4954-bede-716091ba57b8-horizon-secret-key\") pod \"horizon-86c95686f7-d9tjn\" (UID: \"4845d02f-a80d-4954-bede-716091ba57b8\") " pod="openstack/horizon-86c95686f7-d9tjn" Dec 06 16:02:41 crc kubenswrapper[4813]: I1206 16:02:41.117825 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4845d02f-a80d-4954-bede-716091ba57b8-logs\") pod \"horizon-86c95686f7-d9tjn\" (UID: \"4845d02f-a80d-4954-bede-716091ba57b8\") " pod="openstack/horizon-86c95686f7-d9tjn" Dec 06 16:02:41 crc kubenswrapper[4813]: I1206 16:02:41.118530 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4845d02f-a80d-4954-bede-716091ba57b8-config-data\") pod \"horizon-86c95686f7-d9tjn\" (UID: \"4845d02f-a80d-4954-bede-716091ba57b8\") " pod="openstack/horizon-86c95686f7-d9tjn" Dec 06 16:02:41 crc kubenswrapper[4813]: I1206 16:02:41.118681 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4845d02f-a80d-4954-bede-716091ba57b8-scripts\") pod \"horizon-86c95686f7-d9tjn\" (UID: \"4845d02f-a80d-4954-bede-716091ba57b8\") " pod="openstack/horizon-86c95686f7-d9tjn" Dec 06 16:02:41 crc kubenswrapper[4813]: I1206 16:02:41.124579 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4845d02f-a80d-4954-bede-716091ba57b8-horizon-secret-key\") pod \"horizon-86c95686f7-d9tjn\" (UID: \"4845d02f-a80d-4954-bede-716091ba57b8\") " pod="openstack/horizon-86c95686f7-d9tjn" Dec 06 16:02:41 crc kubenswrapper[4813]: I1206 16:02:41.154515 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsqxj\" (UniqueName: \"kubernetes.io/projected/4845d02f-a80d-4954-bede-716091ba57b8-kube-api-access-bsqxj\") pod \"horizon-86c95686f7-d9tjn\" (UID: \"4845d02f-a80d-4954-bede-716091ba57b8\") " pod="openstack/horizon-86c95686f7-d9tjn" Dec 06 16:02:41 crc kubenswrapper[4813]: I1206 16:02:41.256162 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-86c95686f7-d9tjn" Dec 06 16:02:42 crc kubenswrapper[4813]: I1206 16:02:42.516961 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bca3c7d-5b59-41d0-b441-546e4417718c" path="/var/lib/kubelet/pods/4bca3c7d-5b59-41d0-b441-546e4417718c/volumes" Dec 06 16:02:44 crc kubenswrapper[4813]: I1206 16:02:44.389058 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-86c95686f7-d9tjn"] Dec 06 16:02:44 crc kubenswrapper[4813]: I1206 16:02:44.660089 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c4340aaa-8da3-4293-87f2-d49f43ee01bd","Type":"ContainerStarted","Data":"ddb27420ae62899fa3986c3835f7f9f44fb5d00e9f89858ac6a7ec912af9a51f"} Dec 06 16:02:44 crc kubenswrapper[4813]: I1206 16:02:44.662780 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"372958b7-ef0c-448e-8acc-57a1b27a985f","Type":"ContainerStarted","Data":"16b64922c56fc8c8c92c7d2d1584e849542a076ea94fe66abe7f0f797c9d706a"} Dec 06 16:02:44 crc kubenswrapper[4813]: I1206 16:02:44.669894 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" event={"ID":"efca4ca3-8021-4417-83cc-1b26f13f97bd","Type":"ContainerStarted","Data":"2b0d09d239154fc3944e334774798b0da04c92d93b83aee3bbb341b976d5155c"} Dec 06 16:02:44 crc kubenswrapper[4813]: I1206 16:02:44.671137 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" Dec 06 16:02:44 crc kubenswrapper[4813]: I1206 16:02:44.699318 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" podStartSLOduration=7.699236491 podStartE2EDuration="7.699236491s" podCreationTimestamp="2025-12-06 16:02:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:02:44.690030997 +0000 UTC m=+1004.580910573" watchObservedRunningTime="2025-12-06 16:02:44.699236491 +0000 UTC m=+1004.590116067" Dec 06 16:02:45 crc kubenswrapper[4813]: E1206 16:02:45.669640 4813 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5d43f84_cbc2_43a7_877c_56d53b8de4b3.slice/crio-conmon-bddf9da2a88bc9810142b3c40b3de2ff32912f6eb820812c93b8fa810c085281.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5d43f84_cbc2_43a7_877c_56d53b8de4b3.slice/crio-bddf9da2a88bc9810142b3c40b3de2ff32912f6eb820812c93b8fa810c085281.scope\": RecentStats: unable to find data in memory cache]" Dec 06 16:02:45 crc kubenswrapper[4813]: I1206 16:02:45.681468 4813 generic.go:334] "Generic (PLEG): container finished" podID="e5d43f84-cbc2-43a7-877c-56d53b8de4b3" containerID="bddf9da2a88bc9810142b3c40b3de2ff32912f6eb820812c93b8fa810c085281" exitCode=0 Dec 06 16:02:45 crc kubenswrapper[4813]: I1206 16:02:45.681611 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zbvhf" event={"ID":"e5d43f84-cbc2-43a7-877c-56d53b8de4b3","Type":"ContainerDied","Data":"bddf9da2a88bc9810142b3c40b3de2ff32912f6eb820812c93b8fa810c085281"} Dec 06 16:02:45 crc kubenswrapper[4813]: I1206 16:02:45.809242 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7f949586b5-5pzsg"] Dec 06 16:02:45 crc kubenswrapper[4813]: I1206 16:02:45.873103 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-bbc668d58-h8w67"] Dec 06 16:02:45 crc kubenswrapper[4813]: I1206 16:02:45.874799 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-bbc668d58-h8w67" Dec 06 16:02:45 crc kubenswrapper[4813]: I1206 16:02:45.883778 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 06 16:02:45 crc kubenswrapper[4813]: I1206 16:02:45.910313 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-bbc668d58-h8w67"] Dec 06 16:02:45 crc kubenswrapper[4813]: I1206 16:02:45.950874 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-config-data\") pod \"horizon-bbc668d58-h8w67\" (UID: \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\") " pod="openstack/horizon-bbc668d58-h8w67" Dec 06 16:02:45 crc kubenswrapper[4813]: I1206 16:02:45.950948 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-combined-ca-bundle\") pod \"horizon-bbc668d58-h8w67\" (UID: \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\") " pod="openstack/horizon-bbc668d58-h8w67" Dec 06 16:02:45 crc kubenswrapper[4813]: I1206 16:02:45.950979 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ttbz\" (UniqueName: \"kubernetes.io/projected/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-kube-api-access-8ttbz\") pod \"horizon-bbc668d58-h8w67\" (UID: \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\") " pod="openstack/horizon-bbc668d58-h8w67" Dec 06 16:02:45 crc kubenswrapper[4813]: I1206 16:02:45.951039 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-horizon-secret-key\") pod \"horizon-bbc668d58-h8w67\" (UID: \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\") " pod="openstack/horizon-bbc668d58-h8w67" Dec 06 16:02:45 crc kubenswrapper[4813]: I1206 16:02:45.951120 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-horizon-tls-certs\") pod \"horizon-bbc668d58-h8w67\" (UID: \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\") " pod="openstack/horizon-bbc668d58-h8w67" Dec 06 16:02:45 crc kubenswrapper[4813]: I1206 16:02:45.951179 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-logs\") pod \"horizon-bbc668d58-h8w67\" (UID: \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\") " pod="openstack/horizon-bbc668d58-h8w67" Dec 06 16:02:45 crc kubenswrapper[4813]: I1206 16:02:45.951204 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-scripts\") pod \"horizon-bbc668d58-h8w67\" (UID: \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\") " pod="openstack/horizon-bbc668d58-h8w67" Dec 06 16:02:45 crc kubenswrapper[4813]: I1206 16:02:45.962537 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-86c95686f7-d9tjn"] Dec 06 16:02:45 crc kubenswrapper[4813]: I1206 16:02:45.996892 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7fd6f7946b-kmm96"] Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.006850 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7fd6f7946b-kmm96"] Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.006944 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7fd6f7946b-kmm96" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.053709 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-horizon-tls-certs\") pod \"horizon-bbc668d58-h8w67\" (UID: \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\") " pod="openstack/horizon-bbc668d58-h8w67" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.053775 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-logs\") pod \"horizon-bbc668d58-h8w67\" (UID: \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\") " pod="openstack/horizon-bbc668d58-h8w67" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.053804 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-scripts\") pod \"horizon-bbc668d58-h8w67\" (UID: \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\") " pod="openstack/horizon-bbc668d58-h8w67" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.053833 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-config-data\") pod \"horizon-bbc668d58-h8w67\" (UID: \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\") " pod="openstack/horizon-bbc668d58-h8w67" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.053862 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-combined-ca-bundle\") pod \"horizon-bbc668d58-h8w67\" (UID: \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\") " pod="openstack/horizon-bbc668d58-h8w67" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.053887 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ttbz\" (UniqueName: \"kubernetes.io/projected/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-kube-api-access-8ttbz\") pod \"horizon-bbc668d58-h8w67\" (UID: \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\") " pod="openstack/horizon-bbc668d58-h8w67" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.053924 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-horizon-secret-key\") pod \"horizon-bbc668d58-h8w67\" (UID: \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\") " pod="openstack/horizon-bbc668d58-h8w67" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.055445 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-scripts\") pod \"horizon-bbc668d58-h8w67\" (UID: \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\") " pod="openstack/horizon-bbc668d58-h8w67" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.056573 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-logs\") pod \"horizon-bbc668d58-h8w67\" (UID: \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\") " pod="openstack/horizon-bbc668d58-h8w67" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.057626 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-config-data\") pod \"horizon-bbc668d58-h8w67\" (UID: \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\") " pod="openstack/horizon-bbc668d58-h8w67" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.062559 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-horizon-secret-key\") pod \"horizon-bbc668d58-h8w67\" (UID: \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\") " pod="openstack/horizon-bbc668d58-h8w67" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.068049 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-horizon-tls-certs\") pod \"horizon-bbc668d58-h8w67\" (UID: \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\") " pod="openstack/horizon-bbc668d58-h8w67" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.071578 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-combined-ca-bundle\") pod \"horizon-bbc668d58-h8w67\" (UID: \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\") " pod="openstack/horizon-bbc668d58-h8w67" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.081408 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ttbz\" (UniqueName: \"kubernetes.io/projected/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-kube-api-access-8ttbz\") pod \"horizon-bbc668d58-h8w67\" (UID: \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\") " pod="openstack/horizon-bbc668d58-h8w67" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.155529 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/28238998-5d44-4ae9-8d1f-93d56ff18152-scripts\") pod \"horizon-7fd6f7946b-kmm96\" (UID: \"28238998-5d44-4ae9-8d1f-93d56ff18152\") " pod="openstack/horizon-7fd6f7946b-kmm96" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.155570 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/28238998-5d44-4ae9-8d1f-93d56ff18152-config-data\") pod \"horizon-7fd6f7946b-kmm96\" (UID: \"28238998-5d44-4ae9-8d1f-93d56ff18152\") " pod="openstack/horizon-7fd6f7946b-kmm96" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.155634 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28238998-5d44-4ae9-8d1f-93d56ff18152-combined-ca-bundle\") pod \"horizon-7fd6f7946b-kmm96\" (UID: \"28238998-5d44-4ae9-8d1f-93d56ff18152\") " pod="openstack/horizon-7fd6f7946b-kmm96" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.155664 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/28238998-5d44-4ae9-8d1f-93d56ff18152-horizon-secret-key\") pod \"horizon-7fd6f7946b-kmm96\" (UID: \"28238998-5d44-4ae9-8d1f-93d56ff18152\") " pod="openstack/horizon-7fd6f7946b-kmm96" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.155693 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/28238998-5d44-4ae9-8d1f-93d56ff18152-horizon-tls-certs\") pod \"horizon-7fd6f7946b-kmm96\" (UID: \"28238998-5d44-4ae9-8d1f-93d56ff18152\") " pod="openstack/horizon-7fd6f7946b-kmm96" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.155717 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9btsq\" (UniqueName: \"kubernetes.io/projected/28238998-5d44-4ae9-8d1f-93d56ff18152-kube-api-access-9btsq\") pod \"horizon-7fd6f7946b-kmm96\" (UID: \"28238998-5d44-4ae9-8d1f-93d56ff18152\") " pod="openstack/horizon-7fd6f7946b-kmm96" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.155756 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28238998-5d44-4ae9-8d1f-93d56ff18152-logs\") pod \"horizon-7fd6f7946b-kmm96\" (UID: \"28238998-5d44-4ae9-8d1f-93d56ff18152\") " pod="openstack/horizon-7fd6f7946b-kmm96" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.230153 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-bbc668d58-h8w67" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.257226 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28238998-5d44-4ae9-8d1f-93d56ff18152-logs\") pod \"horizon-7fd6f7946b-kmm96\" (UID: \"28238998-5d44-4ae9-8d1f-93d56ff18152\") " pod="openstack/horizon-7fd6f7946b-kmm96" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.257324 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/28238998-5d44-4ae9-8d1f-93d56ff18152-scripts\") pod \"horizon-7fd6f7946b-kmm96\" (UID: \"28238998-5d44-4ae9-8d1f-93d56ff18152\") " pod="openstack/horizon-7fd6f7946b-kmm96" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.257349 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/28238998-5d44-4ae9-8d1f-93d56ff18152-config-data\") pod \"horizon-7fd6f7946b-kmm96\" (UID: \"28238998-5d44-4ae9-8d1f-93d56ff18152\") " pod="openstack/horizon-7fd6f7946b-kmm96" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.257514 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28238998-5d44-4ae9-8d1f-93d56ff18152-combined-ca-bundle\") pod \"horizon-7fd6f7946b-kmm96\" (UID: \"28238998-5d44-4ae9-8d1f-93d56ff18152\") " pod="openstack/horizon-7fd6f7946b-kmm96" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.259107 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/28238998-5d44-4ae9-8d1f-93d56ff18152-horizon-secret-key\") pod \"horizon-7fd6f7946b-kmm96\" (UID: \"28238998-5d44-4ae9-8d1f-93d56ff18152\") " pod="openstack/horizon-7fd6f7946b-kmm96" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.257807 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28238998-5d44-4ae9-8d1f-93d56ff18152-logs\") pod \"horizon-7fd6f7946b-kmm96\" (UID: \"28238998-5d44-4ae9-8d1f-93d56ff18152\") " pod="openstack/horizon-7fd6f7946b-kmm96" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.259043 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/28238998-5d44-4ae9-8d1f-93d56ff18152-config-data\") pod \"horizon-7fd6f7946b-kmm96\" (UID: \"28238998-5d44-4ae9-8d1f-93d56ff18152\") " pod="openstack/horizon-7fd6f7946b-kmm96" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.258172 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/28238998-5d44-4ae9-8d1f-93d56ff18152-scripts\") pod \"horizon-7fd6f7946b-kmm96\" (UID: \"28238998-5d44-4ae9-8d1f-93d56ff18152\") " pod="openstack/horizon-7fd6f7946b-kmm96" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.259277 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/28238998-5d44-4ae9-8d1f-93d56ff18152-horizon-tls-certs\") pod \"horizon-7fd6f7946b-kmm96\" (UID: \"28238998-5d44-4ae9-8d1f-93d56ff18152\") " pod="openstack/horizon-7fd6f7946b-kmm96" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.259309 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9btsq\" (UniqueName: \"kubernetes.io/projected/28238998-5d44-4ae9-8d1f-93d56ff18152-kube-api-access-9btsq\") pod \"horizon-7fd6f7946b-kmm96\" (UID: \"28238998-5d44-4ae9-8d1f-93d56ff18152\") " pod="openstack/horizon-7fd6f7946b-kmm96" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.262888 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28238998-5d44-4ae9-8d1f-93d56ff18152-combined-ca-bundle\") pod \"horizon-7fd6f7946b-kmm96\" (UID: \"28238998-5d44-4ae9-8d1f-93d56ff18152\") " pod="openstack/horizon-7fd6f7946b-kmm96" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.265566 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/28238998-5d44-4ae9-8d1f-93d56ff18152-horizon-tls-certs\") pod \"horizon-7fd6f7946b-kmm96\" (UID: \"28238998-5d44-4ae9-8d1f-93d56ff18152\") " pod="openstack/horizon-7fd6f7946b-kmm96" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.274595 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/28238998-5d44-4ae9-8d1f-93d56ff18152-horizon-secret-key\") pod \"horizon-7fd6f7946b-kmm96\" (UID: \"28238998-5d44-4ae9-8d1f-93d56ff18152\") " pod="openstack/horizon-7fd6f7946b-kmm96" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.275504 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9btsq\" (UniqueName: \"kubernetes.io/projected/28238998-5d44-4ae9-8d1f-93d56ff18152-kube-api-access-9btsq\") pod \"horizon-7fd6f7946b-kmm96\" (UID: \"28238998-5d44-4ae9-8d1f-93d56ff18152\") " pod="openstack/horizon-7fd6f7946b-kmm96" Dec 06 16:02:46 crc kubenswrapper[4813]: I1206 16:02:46.343805 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7fd6f7946b-kmm96" Dec 06 16:02:49 crc kubenswrapper[4813]: I1206 16:02:49.427936 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:02:49 crc kubenswrapper[4813]: I1206 16:02:49.428502 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:02:50 crc kubenswrapper[4813]: I1206 16:02:50.006929 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zbvhf" Dec 06 16:02:50 crc kubenswrapper[4813]: I1206 16:02:50.123235 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-config-data\") pod \"e5d43f84-cbc2-43a7-877c-56d53b8de4b3\" (UID: \"e5d43f84-cbc2-43a7-877c-56d53b8de4b3\") " Dec 06 16:02:50 crc kubenswrapper[4813]: I1206 16:02:50.123300 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-credential-keys\") pod \"e5d43f84-cbc2-43a7-877c-56d53b8de4b3\" (UID: \"e5d43f84-cbc2-43a7-877c-56d53b8de4b3\") " Dec 06 16:02:50 crc kubenswrapper[4813]: I1206 16:02:50.123411 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-scripts\") pod \"e5d43f84-cbc2-43a7-877c-56d53b8de4b3\" (UID: \"e5d43f84-cbc2-43a7-877c-56d53b8de4b3\") " Dec 06 16:02:50 crc kubenswrapper[4813]: I1206 16:02:50.123447 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-combined-ca-bundle\") pod \"e5d43f84-cbc2-43a7-877c-56d53b8de4b3\" (UID: \"e5d43f84-cbc2-43a7-877c-56d53b8de4b3\") " Dec 06 16:02:50 crc kubenswrapper[4813]: I1206 16:02:50.123524 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-fernet-keys\") pod \"e5d43f84-cbc2-43a7-877c-56d53b8de4b3\" (UID: \"e5d43f84-cbc2-43a7-877c-56d53b8de4b3\") " Dec 06 16:02:50 crc kubenswrapper[4813]: I1206 16:02:50.123563 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsrk8\" (UniqueName: \"kubernetes.io/projected/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-kube-api-access-gsrk8\") pod \"e5d43f84-cbc2-43a7-877c-56d53b8de4b3\" (UID: \"e5d43f84-cbc2-43a7-877c-56d53b8de4b3\") " Dec 06 16:02:50 crc kubenswrapper[4813]: I1206 16:02:50.128634 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-scripts" (OuterVolumeSpecName: "scripts") pod "e5d43f84-cbc2-43a7-877c-56d53b8de4b3" (UID: "e5d43f84-cbc2-43a7-877c-56d53b8de4b3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:02:50 crc kubenswrapper[4813]: I1206 16:02:50.131030 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "e5d43f84-cbc2-43a7-877c-56d53b8de4b3" (UID: "e5d43f84-cbc2-43a7-877c-56d53b8de4b3"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:02:50 crc kubenswrapper[4813]: I1206 16:02:50.132549 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-kube-api-access-gsrk8" (OuterVolumeSpecName: "kube-api-access-gsrk8") pod "e5d43f84-cbc2-43a7-877c-56d53b8de4b3" (UID: "e5d43f84-cbc2-43a7-877c-56d53b8de4b3"). InnerVolumeSpecName "kube-api-access-gsrk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:02:50 crc kubenswrapper[4813]: I1206 16:02:50.133895 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "e5d43f84-cbc2-43a7-877c-56d53b8de4b3" (UID: "e5d43f84-cbc2-43a7-877c-56d53b8de4b3"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:02:50 crc kubenswrapper[4813]: I1206 16:02:50.160457 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e5d43f84-cbc2-43a7-877c-56d53b8de4b3" (UID: "e5d43f84-cbc2-43a7-877c-56d53b8de4b3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:02:50 crc kubenswrapper[4813]: I1206 16:02:50.166142 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-config-data" (OuterVolumeSpecName: "config-data") pod "e5d43f84-cbc2-43a7-877c-56d53b8de4b3" (UID: "e5d43f84-cbc2-43a7-877c-56d53b8de4b3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:02:50 crc kubenswrapper[4813]: I1206 16:02:50.226570 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:50 crc kubenswrapper[4813]: I1206 16:02:50.226600 4813 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:50 crc kubenswrapper[4813]: I1206 16:02:50.226612 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:50 crc kubenswrapper[4813]: I1206 16:02:50.226621 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:50 crc kubenswrapper[4813]: I1206 16:02:50.226631 4813 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:50 crc kubenswrapper[4813]: I1206 16:02:50.226639 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsrk8\" (UniqueName: \"kubernetes.io/projected/e5d43f84-cbc2-43a7-877c-56d53b8de4b3-kube-api-access-gsrk8\") on node \"crc\" DevicePath \"\"" Dec 06 16:02:50 crc kubenswrapper[4813]: I1206 16:02:50.758320 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zbvhf" event={"ID":"e5d43f84-cbc2-43a7-877c-56d53b8de4b3","Type":"ContainerDied","Data":"b48a4a245b0a4a48d62913b12ddba81cb7c75aea32427ae2dd8da1ec0df66aad"} Dec 06 16:02:50 crc kubenswrapper[4813]: I1206 16:02:50.758967 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b48a4a245b0a4a48d62913b12ddba81cb7c75aea32427ae2dd8da1ec0df66aad" Dec 06 16:02:50 crc kubenswrapper[4813]: I1206 16:02:50.758377 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zbvhf" Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.100105 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-zbvhf"] Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.112894 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-zbvhf"] Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.193734 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-6qjfq"] Dec 06 16:02:51 crc kubenswrapper[4813]: E1206 16:02:51.194372 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5d43f84-cbc2-43a7-877c-56d53b8de4b3" containerName="keystone-bootstrap" Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.197572 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5d43f84-cbc2-43a7-877c-56d53b8de4b3" containerName="keystone-bootstrap" Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.197948 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5d43f84-cbc2-43a7-877c-56d53b8de4b3" containerName="keystone-bootstrap" Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.198561 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6qjfq" Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.214095 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.214429 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.214614 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.214750 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.215702 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fmvk2" Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.238389 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-6qjfq"] Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.254923 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-455vp\" (UniqueName: \"kubernetes.io/projected/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-kube-api-access-455vp\") pod \"keystone-bootstrap-6qjfq\" (UID: \"67cf9ebe-1a58-4e50-9c10-f6f7933e781f\") " pod="openstack/keystone-bootstrap-6qjfq" Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.254991 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-scripts\") pod \"keystone-bootstrap-6qjfq\" (UID: \"67cf9ebe-1a58-4e50-9c10-f6f7933e781f\") " pod="openstack/keystone-bootstrap-6qjfq" Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.255042 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-credential-keys\") pod \"keystone-bootstrap-6qjfq\" (UID: \"67cf9ebe-1a58-4e50-9c10-f6f7933e781f\") " pod="openstack/keystone-bootstrap-6qjfq" Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.255060 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-fernet-keys\") pod \"keystone-bootstrap-6qjfq\" (UID: \"67cf9ebe-1a58-4e50-9c10-f6f7933e781f\") " pod="openstack/keystone-bootstrap-6qjfq" Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.255100 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-combined-ca-bundle\") pod \"keystone-bootstrap-6qjfq\" (UID: \"67cf9ebe-1a58-4e50-9c10-f6f7933e781f\") " pod="openstack/keystone-bootstrap-6qjfq" Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.255170 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-config-data\") pod \"keystone-bootstrap-6qjfq\" (UID: \"67cf9ebe-1a58-4e50-9c10-f6f7933e781f\") " pod="openstack/keystone-bootstrap-6qjfq" Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.356944 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-combined-ca-bundle\") pod \"keystone-bootstrap-6qjfq\" (UID: \"67cf9ebe-1a58-4e50-9c10-f6f7933e781f\") " pod="openstack/keystone-bootstrap-6qjfq" Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.357039 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-config-data\") pod \"keystone-bootstrap-6qjfq\" (UID: \"67cf9ebe-1a58-4e50-9c10-f6f7933e781f\") " pod="openstack/keystone-bootstrap-6qjfq" Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.357079 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-455vp\" (UniqueName: \"kubernetes.io/projected/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-kube-api-access-455vp\") pod \"keystone-bootstrap-6qjfq\" (UID: \"67cf9ebe-1a58-4e50-9c10-f6f7933e781f\") " pod="openstack/keystone-bootstrap-6qjfq" Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.357096 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-scripts\") pod \"keystone-bootstrap-6qjfq\" (UID: \"67cf9ebe-1a58-4e50-9c10-f6f7933e781f\") " pod="openstack/keystone-bootstrap-6qjfq" Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.357137 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-credential-keys\") pod \"keystone-bootstrap-6qjfq\" (UID: \"67cf9ebe-1a58-4e50-9c10-f6f7933e781f\") " pod="openstack/keystone-bootstrap-6qjfq" Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.357151 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-fernet-keys\") pod \"keystone-bootstrap-6qjfq\" (UID: \"67cf9ebe-1a58-4e50-9c10-f6f7933e781f\") " pod="openstack/keystone-bootstrap-6qjfq" Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.366301 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-fernet-keys\") pod \"keystone-bootstrap-6qjfq\" (UID: \"67cf9ebe-1a58-4e50-9c10-f6f7933e781f\") " pod="openstack/keystone-bootstrap-6qjfq" Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.369909 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-combined-ca-bundle\") pod \"keystone-bootstrap-6qjfq\" (UID: \"67cf9ebe-1a58-4e50-9c10-f6f7933e781f\") " pod="openstack/keystone-bootstrap-6qjfq" Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.376482 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-scripts\") pod \"keystone-bootstrap-6qjfq\" (UID: \"67cf9ebe-1a58-4e50-9c10-f6f7933e781f\") " pod="openstack/keystone-bootstrap-6qjfq" Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.376696 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-455vp\" (UniqueName: \"kubernetes.io/projected/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-kube-api-access-455vp\") pod \"keystone-bootstrap-6qjfq\" (UID: \"67cf9ebe-1a58-4e50-9c10-f6f7933e781f\") " pod="openstack/keystone-bootstrap-6qjfq" Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.379833 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-credential-keys\") pod \"keystone-bootstrap-6qjfq\" (UID: \"67cf9ebe-1a58-4e50-9c10-f6f7933e781f\") " pod="openstack/keystone-bootstrap-6qjfq" Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.383662 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-config-data\") pod \"keystone-bootstrap-6qjfq\" (UID: \"67cf9ebe-1a58-4e50-9c10-f6f7933e781f\") " pod="openstack/keystone-bootstrap-6qjfq" Dec 06 16:02:51 crc kubenswrapper[4813]: I1206 16:02:51.520569 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6qjfq" Dec 06 16:02:52 crc kubenswrapper[4813]: I1206 16:02:52.496447 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5d43f84-cbc2-43a7-877c-56d53b8de4b3" path="/var/lib/kubelet/pods/e5d43f84-cbc2-43a7-877c-56d53b8de4b3/volumes" Dec 06 16:02:54 crc kubenswrapper[4813]: I1206 16:02:53.631838 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" Dec 06 16:02:54 crc kubenswrapper[4813]: I1206 16:02:54.123077 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-tbd65"] Dec 06 16:02:54 crc kubenswrapper[4813]: I1206 16:02:54.123307 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" podUID="274211c2-377d-4164-be60-ca42e8167849" containerName="dnsmasq-dns" containerID="cri-o://554a49704af221c2f375a789f3ae52a335b47c491aacfa5eacdbc02f838c3401" gracePeriod=10 Dec 06 16:02:55 crc kubenswrapper[4813]: I1206 16:02:55.078605 4813 generic.go:334] "Generic (PLEG): container finished" podID="274211c2-377d-4164-be60-ca42e8167849" containerID="554a49704af221c2f375a789f3ae52a335b47c491aacfa5eacdbc02f838c3401" exitCode=0 Dec 06 16:02:55 crc kubenswrapper[4813]: I1206 16:02:55.078673 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" event={"ID":"274211c2-377d-4164-be60-ca42e8167849","Type":"ContainerDied","Data":"554a49704af221c2f375a789f3ae52a335b47c491aacfa5eacdbc02f838c3401"} Dec 06 16:02:55 crc kubenswrapper[4813]: I1206 16:02:55.292443 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" podUID="274211c2-377d-4164-be60-ca42e8167849" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: connect: connection refused" Dec 06 16:02:59 crc kubenswrapper[4813]: E1206 16:02:59.748659 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 06 16:02:59 crc kubenswrapper[4813]: E1206 16:02:59.749649 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n8h5bchfhd5hc4hc9hb4hc4h5bbh597h57bh88h555h88h546hc9hddh5b6h5b7h545hb7h9bhbfh58fhfhf6h54ch66chch5f5h66dhf8q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2cn8f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-6b588dd7f7-bbddl_openstack(ef70d3a8-1fae-4950-b8c8-e12b48b7dac1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 16:02:59 crc kubenswrapper[4813]: E1206 16:02:59.805473 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-6b588dd7f7-bbddl" podUID="ef70d3a8-1fae-4950-b8c8-e12b48b7dac1" Dec 06 16:03:00 crc kubenswrapper[4813]: I1206 16:03:00.122386 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-86c95686f7-d9tjn" event={"ID":"4845d02f-a80d-4954-bede-716091ba57b8","Type":"ContainerStarted","Data":"3bb2356ba7e868fb83f42a0a4ca8718b964a8862f2b0bba7261330de4fcf42c0"} Dec 06 16:03:00 crc kubenswrapper[4813]: I1206 16:03:00.292034 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" podUID="274211c2-377d-4164-be60-ca42e8167849" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: connect: connection refused" Dec 06 16:03:01 crc kubenswrapper[4813]: E1206 16:03:01.653829 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 06 16:03:01 crc kubenswrapper[4813]: E1206 16:03:01.654000 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n98h569hf8hf6h6h665h675h647hc4h57bh77hf5h684h588hd5h94h97h89h5bfhb7h595hdfh54bh5bdhbdh68dh65bhfbh9fh57ch95h566q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-95zr6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7f949586b5-5pzsg_openstack(b697e541-1bb3-4619-8cc6-ce548717c741): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 16:03:01 crc kubenswrapper[4813]: E1206 16:03:01.657367 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7f949586b5-5pzsg" podUID="b697e541-1bb3-4619-8cc6-ce548717c741" Dec 06 16:03:06 crc kubenswrapper[4813]: I1206 16:03:06.167787 4813 generic.go:334] "Generic (PLEG): container finished" podID="72532d82-6482-41ac-a461-c792c134dee3" containerID="c421e9528a587793ae5ac6d5f50537230cf1f38bb2b5734e64d67f19d34e4772" exitCode=0 Dec 06 16:03:06 crc kubenswrapper[4813]: I1206 16:03:06.167873 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-5zx65" event={"ID":"72532d82-6482-41ac-a461-c792c134dee3","Type":"ContainerDied","Data":"c421e9528a587793ae5ac6d5f50537230cf1f38bb2b5734e64d67f19d34e4772"} Dec 06 16:03:10 crc kubenswrapper[4813]: I1206 16:03:10.291352 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" podUID="274211c2-377d-4164-be60-ca42e8167849" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: i/o timeout" Dec 06 16:03:10 crc kubenswrapper[4813]: I1206 16:03:10.291911 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.231879 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6b588dd7f7-bbddl" event={"ID":"ef70d3a8-1fae-4950-b8c8-e12b48b7dac1","Type":"ContainerDied","Data":"d899bfdaf7bd3ace3f0c065abdf4067ae59ddcd2d40e92edc122c4a203fd91cc"} Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.232652 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d899bfdaf7bd3ace3f0c065abdf4067ae59ddcd2d40e92edc122c4a203fd91cc" Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.306011 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6b588dd7f7-bbddl" Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.413388 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1-scripts\") pod \"ef70d3a8-1fae-4950-b8c8-e12b48b7dac1\" (UID: \"ef70d3a8-1fae-4950-b8c8-e12b48b7dac1\") " Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.413740 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1-horizon-secret-key\") pod \"ef70d3a8-1fae-4950-b8c8-e12b48b7dac1\" (UID: \"ef70d3a8-1fae-4950-b8c8-e12b48b7dac1\") " Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.413805 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1-logs\") pod \"ef70d3a8-1fae-4950-b8c8-e12b48b7dac1\" (UID: \"ef70d3a8-1fae-4950-b8c8-e12b48b7dac1\") " Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.413860 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1-config-data\") pod \"ef70d3a8-1fae-4950-b8c8-e12b48b7dac1\" (UID: \"ef70d3a8-1fae-4950-b8c8-e12b48b7dac1\") " Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.413881 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cn8f\" (UniqueName: \"kubernetes.io/projected/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1-kube-api-access-2cn8f\") pod \"ef70d3a8-1fae-4950-b8c8-e12b48b7dac1\" (UID: \"ef70d3a8-1fae-4950-b8c8-e12b48b7dac1\") " Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.413975 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1-scripts" (OuterVolumeSpecName: "scripts") pod "ef70d3a8-1fae-4950-b8c8-e12b48b7dac1" (UID: "ef70d3a8-1fae-4950-b8c8-e12b48b7dac1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.414235 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.414238 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1-logs" (OuterVolumeSpecName: "logs") pod "ef70d3a8-1fae-4950-b8c8-e12b48b7dac1" (UID: "ef70d3a8-1fae-4950-b8c8-e12b48b7dac1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.414504 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1-config-data" (OuterVolumeSpecName: "config-data") pod "ef70d3a8-1fae-4950-b8c8-e12b48b7dac1" (UID: "ef70d3a8-1fae-4950-b8c8-e12b48b7dac1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.439553 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "ef70d3a8-1fae-4950-b8c8-e12b48b7dac1" (UID: "ef70d3a8-1fae-4950-b8c8-e12b48b7dac1"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.439910 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1-kube-api-access-2cn8f" (OuterVolumeSpecName: "kube-api-access-2cn8f") pod "ef70d3a8-1fae-4950-b8c8-e12b48b7dac1" (UID: "ef70d3a8-1fae-4950-b8c8-e12b48b7dac1"). InnerVolumeSpecName "kube-api-access-2cn8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.515396 4813 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.515420 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1-logs\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.515664 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.515697 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cn8f\" (UniqueName: \"kubernetes.io/projected/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1-kube-api-access-2cn8f\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.690552 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7fd6f7946b-kmm96"] Dec 06 16:03:13 crc kubenswrapper[4813]: E1206 16:03:13.717721 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Dec 06 16:03:13 crc kubenswrapper[4813]: E1206 16:03:13.717967 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n665h587h567hbbh94h686h9bh5bch676h7ch5b8h54bh68bh6h544h5c6h554h89hfdh689h668h5fbh665h5c4h565h88hd4hf4h5cch55dhc4h5f8q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k6tkg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(f539e8ab-300c-426b-98e7-f9e87a6787b5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.805667 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.811913 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-5zx65" Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.826646 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f949586b5-5pzsg" Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.922019 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/274211c2-377d-4164-be60-ca42e8167849-dns-svc\") pod \"274211c2-377d-4164-be60-ca42e8167849\" (UID: \"274211c2-377d-4164-be60-ca42e8167849\") " Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.922402 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95zr6\" (UniqueName: \"kubernetes.io/projected/b697e541-1bb3-4619-8cc6-ce548717c741-kube-api-access-95zr6\") pod \"b697e541-1bb3-4619-8cc6-ce548717c741\" (UID: \"b697e541-1bb3-4619-8cc6-ce548717c741\") " Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.922572 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lclqs\" (UniqueName: \"kubernetes.io/projected/274211c2-377d-4164-be60-ca42e8167849-kube-api-access-lclqs\") pod \"274211c2-377d-4164-be60-ca42e8167849\" (UID: \"274211c2-377d-4164-be60-ca42e8167849\") " Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.922621 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72532d82-6482-41ac-a461-c792c134dee3-combined-ca-bundle\") pod \"72532d82-6482-41ac-a461-c792c134dee3\" (UID: \"72532d82-6482-41ac-a461-c792c134dee3\") " Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.922657 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/274211c2-377d-4164-be60-ca42e8167849-config\") pod \"274211c2-377d-4164-be60-ca42e8167849\" (UID: \"274211c2-377d-4164-be60-ca42e8167849\") " Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.922691 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/274211c2-377d-4164-be60-ca42e8167849-ovsdbserver-nb\") pod \"274211c2-377d-4164-be60-ca42e8167849\" (UID: \"274211c2-377d-4164-be60-ca42e8167849\") " Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.922736 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/274211c2-377d-4164-be60-ca42e8167849-dns-swift-storage-0\") pod \"274211c2-377d-4164-be60-ca42e8167849\" (UID: \"274211c2-377d-4164-be60-ca42e8167849\") " Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.922765 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/72532d82-6482-41ac-a461-c792c134dee3-config\") pod \"72532d82-6482-41ac-a461-c792c134dee3\" (UID: \"72532d82-6482-41ac-a461-c792c134dee3\") " Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.922792 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b697e541-1bb3-4619-8cc6-ce548717c741-horizon-secret-key\") pod \"b697e541-1bb3-4619-8cc6-ce548717c741\" (UID: \"b697e541-1bb3-4619-8cc6-ce548717c741\") " Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.922818 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6gk8\" (UniqueName: \"kubernetes.io/projected/72532d82-6482-41ac-a461-c792c134dee3-kube-api-access-l6gk8\") pod \"72532d82-6482-41ac-a461-c792c134dee3\" (UID: \"72532d82-6482-41ac-a461-c792c134dee3\") " Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.922851 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b697e541-1bb3-4619-8cc6-ce548717c741-config-data\") pod \"b697e541-1bb3-4619-8cc6-ce548717c741\" (UID: \"b697e541-1bb3-4619-8cc6-ce548717c741\") " Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.922885 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b697e541-1bb3-4619-8cc6-ce548717c741-logs\") pod \"b697e541-1bb3-4619-8cc6-ce548717c741\" (UID: \"b697e541-1bb3-4619-8cc6-ce548717c741\") " Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.922918 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/274211c2-377d-4164-be60-ca42e8167849-ovsdbserver-sb\") pod \"274211c2-377d-4164-be60-ca42e8167849\" (UID: \"274211c2-377d-4164-be60-ca42e8167849\") " Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.922967 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b697e541-1bb3-4619-8cc6-ce548717c741-scripts\") pod \"b697e541-1bb3-4619-8cc6-ce548717c741\" (UID: \"b697e541-1bb3-4619-8cc6-ce548717c741\") " Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.923780 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b697e541-1bb3-4619-8cc6-ce548717c741-scripts" (OuterVolumeSpecName: "scripts") pod "b697e541-1bb3-4619-8cc6-ce548717c741" (UID: "b697e541-1bb3-4619-8cc6-ce548717c741"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.923820 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b697e541-1bb3-4619-8cc6-ce548717c741-config-data" (OuterVolumeSpecName: "config-data") pod "b697e541-1bb3-4619-8cc6-ce548717c741" (UID: "b697e541-1bb3-4619-8cc6-ce548717c741"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.924209 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b697e541-1bb3-4619-8cc6-ce548717c741-logs" (OuterVolumeSpecName: "logs") pod "b697e541-1bb3-4619-8cc6-ce548717c741" (UID: "b697e541-1bb3-4619-8cc6-ce548717c741"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.927502 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b697e541-1bb3-4619-8cc6-ce548717c741-kube-api-access-95zr6" (OuterVolumeSpecName: "kube-api-access-95zr6") pod "b697e541-1bb3-4619-8cc6-ce548717c741" (UID: "b697e541-1bb3-4619-8cc6-ce548717c741"). InnerVolumeSpecName "kube-api-access-95zr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.927904 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b697e541-1bb3-4619-8cc6-ce548717c741-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "b697e541-1bb3-4619-8cc6-ce548717c741" (UID: "b697e541-1bb3-4619-8cc6-ce548717c741"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.928229 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/274211c2-377d-4164-be60-ca42e8167849-kube-api-access-lclqs" (OuterVolumeSpecName: "kube-api-access-lclqs") pod "274211c2-377d-4164-be60-ca42e8167849" (UID: "274211c2-377d-4164-be60-ca42e8167849"). InnerVolumeSpecName "kube-api-access-lclqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.933231 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72532d82-6482-41ac-a461-c792c134dee3-kube-api-access-l6gk8" (OuterVolumeSpecName: "kube-api-access-l6gk8") pod "72532d82-6482-41ac-a461-c792c134dee3" (UID: "72532d82-6482-41ac-a461-c792c134dee3"). InnerVolumeSpecName "kube-api-access-l6gk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.959519 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72532d82-6482-41ac-a461-c792c134dee3-config" (OuterVolumeSpecName: "config") pod "72532d82-6482-41ac-a461-c792c134dee3" (UID: "72532d82-6482-41ac-a461-c792c134dee3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.960024 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72532d82-6482-41ac-a461-c792c134dee3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "72532d82-6482-41ac-a461-c792c134dee3" (UID: "72532d82-6482-41ac-a461-c792c134dee3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.975587 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/274211c2-377d-4164-be60-ca42e8167849-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "274211c2-377d-4164-be60-ca42e8167849" (UID: "274211c2-377d-4164-be60-ca42e8167849"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.975626 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/274211c2-377d-4164-be60-ca42e8167849-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "274211c2-377d-4164-be60-ca42e8167849" (UID: "274211c2-377d-4164-be60-ca42e8167849"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.978093 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/274211c2-377d-4164-be60-ca42e8167849-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "274211c2-377d-4164-be60-ca42e8167849" (UID: "274211c2-377d-4164-be60-ca42e8167849"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.980653 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/274211c2-377d-4164-be60-ca42e8167849-config" (OuterVolumeSpecName: "config") pod "274211c2-377d-4164-be60-ca42e8167849" (UID: "274211c2-377d-4164-be60-ca42e8167849"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:03:13 crc kubenswrapper[4813]: I1206 16:03:13.983780 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/274211c2-377d-4164-be60-ca42e8167849-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "274211c2-377d-4164-be60-ca42e8167849" (UID: "274211c2-377d-4164-be60-ca42e8167849"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.024652 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72532d82-6482-41ac-a461-c792c134dee3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.024688 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/274211c2-377d-4164-be60-ca42e8167849-config\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.024698 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/274211c2-377d-4164-be60-ca42e8167849-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.024707 4813 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/274211c2-377d-4164-be60-ca42e8167849-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.024781 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/72532d82-6482-41ac-a461-c792c134dee3-config\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.024791 4813 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b697e541-1bb3-4619-8cc6-ce548717c741-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.024799 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6gk8\" (UniqueName: \"kubernetes.io/projected/72532d82-6482-41ac-a461-c792c134dee3-kube-api-access-l6gk8\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.024810 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b697e541-1bb3-4619-8cc6-ce548717c741-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.024818 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b697e541-1bb3-4619-8cc6-ce548717c741-logs\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.024827 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/274211c2-377d-4164-be60-ca42e8167849-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.024834 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b697e541-1bb3-4619-8cc6-ce548717c741-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.024842 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/274211c2-377d-4164-be60-ca42e8167849-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.024849 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95zr6\" (UniqueName: \"kubernetes.io/projected/b697e541-1bb3-4619-8cc6-ce548717c741-kube-api-access-95zr6\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.024857 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lclqs\" (UniqueName: \"kubernetes.io/projected/274211c2-377d-4164-be60-ca42e8167849-kube-api-access-lclqs\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.241212 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" event={"ID":"274211c2-377d-4164-be60-ca42e8167849","Type":"ContainerDied","Data":"72cbdf2a26fc9e3245239a0b45e78308483faa2232e2b22236f03bc54e32b645"} Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.241278 4813 scope.go:117] "RemoveContainer" containerID="554a49704af221c2f375a789f3ae52a335b47c491aacfa5eacdbc02f838c3401" Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.241392 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.248010 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f949586b5-5pzsg" event={"ID":"b697e541-1bb3-4619-8cc6-ce548717c741","Type":"ContainerDied","Data":"b88c6e92ef8019f49b583d9c183ae296a31a09108e33269f9efa157e3fdeddb1"} Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.248059 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f949586b5-5pzsg" Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.251614 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-5zx65" Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.251622 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6b588dd7f7-bbddl" Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.251661 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-5zx65" event={"ID":"72532d82-6482-41ac-a461-c792c134dee3","Type":"ContainerDied","Data":"4ba2406e5efc299c12ecae4c70e3ed7df3038b736e2d464e2c419bdc3392bcb2"} Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.251685 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ba2406e5efc299c12ecae4c70e3ed7df3038b736e2d464e2c419bdc3392bcb2" Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.350993 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-tbd65"] Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.358770 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-tbd65"] Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.384378 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6b588dd7f7-bbddl"] Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.390384 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6b588dd7f7-bbddl"] Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.401444 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7f949586b5-5pzsg"] Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.406999 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7f949586b5-5pzsg"] Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.497351 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="274211c2-377d-4164-be60-ca42e8167849" path="/var/lib/kubelet/pods/274211c2-377d-4164-be60-ca42e8167849/volumes" Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.497971 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b697e541-1bb3-4619-8cc6-ce548717c741" path="/var/lib/kubelet/pods/b697e541-1bb3-4619-8cc6-ce548717c741/volumes" Dec 06 16:03:14 crc kubenswrapper[4813]: I1206 16:03:14.498411 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef70d3a8-1fae-4950-b8c8-e12b48b7dac1" path="/var/lib/kubelet/pods/ef70d3a8-1fae-4950-b8c8-e12b48b7dac1/volumes" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.028421 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-qdtmw"] Dec 06 16:03:15 crc kubenswrapper[4813]: E1206 16:03:15.028933 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72532d82-6482-41ac-a461-c792c134dee3" containerName="neutron-db-sync" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.028949 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="72532d82-6482-41ac-a461-c792c134dee3" containerName="neutron-db-sync" Dec 06 16:03:15 crc kubenswrapper[4813]: E1206 16:03:15.028960 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="274211c2-377d-4164-be60-ca42e8167849" containerName="dnsmasq-dns" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.028967 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="274211c2-377d-4164-be60-ca42e8167849" containerName="dnsmasq-dns" Dec 06 16:03:15 crc kubenswrapper[4813]: E1206 16:03:15.028978 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="274211c2-377d-4164-be60-ca42e8167849" containerName="init" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.028984 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="274211c2-377d-4164-be60-ca42e8167849" containerName="init" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.029144 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="72532d82-6482-41ac-a461-c792c134dee3" containerName="neutron-db-sync" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.029172 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="274211c2-377d-4164-be60-ca42e8167849" containerName="dnsmasq-dns" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.032173 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.053066 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-qdtmw"] Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.172164 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90124024-daf7-4241-a787-2840359fe074-dns-svc\") pod \"dnsmasq-dns-6b7b667979-qdtmw\" (UID: \"90124024-daf7-4241-a787-2840359fe074\") " pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.172225 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90124024-daf7-4241-a787-2840359fe074-config\") pod \"dnsmasq-dns-6b7b667979-qdtmw\" (UID: \"90124024-daf7-4241-a787-2840359fe074\") " pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.172582 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/90124024-daf7-4241-a787-2840359fe074-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-qdtmw\" (UID: \"90124024-daf7-4241-a787-2840359fe074\") " pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.172630 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/90124024-daf7-4241-a787-2840359fe074-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-qdtmw\" (UID: \"90124024-daf7-4241-a787-2840359fe074\") " pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.172688 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/90124024-daf7-4241-a787-2840359fe074-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-qdtmw\" (UID: \"90124024-daf7-4241-a787-2840359fe074\") " pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.172803 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2tb4\" (UniqueName: \"kubernetes.io/projected/90124024-daf7-4241-a787-2840359fe074-kube-api-access-c2tb4\") pod \"dnsmasq-dns-6b7b667979-qdtmw\" (UID: \"90124024-daf7-4241-a787-2840359fe074\") " pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.217407 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7c8fc98b5b-dwkdc"] Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.220720 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7c8fc98b5b-dwkdc" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.224694 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-qmg7c" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.225364 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.225370 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.226613 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.247731 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7c8fc98b5b-dwkdc"] Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.274345 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90124024-daf7-4241-a787-2840359fe074-config\") pod \"dnsmasq-dns-6b7b667979-qdtmw\" (UID: \"90124024-daf7-4241-a787-2840359fe074\") " pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.274406 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/90124024-daf7-4241-a787-2840359fe074-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-qdtmw\" (UID: \"90124024-daf7-4241-a787-2840359fe074\") " pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.274439 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/90124024-daf7-4241-a787-2840359fe074-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-qdtmw\" (UID: \"90124024-daf7-4241-a787-2840359fe074\") " pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.274481 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/90124024-daf7-4241-a787-2840359fe074-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-qdtmw\" (UID: \"90124024-daf7-4241-a787-2840359fe074\") " pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.274505 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2tb4\" (UniqueName: \"kubernetes.io/projected/90124024-daf7-4241-a787-2840359fe074-kube-api-access-c2tb4\") pod \"dnsmasq-dns-6b7b667979-qdtmw\" (UID: \"90124024-daf7-4241-a787-2840359fe074\") " pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.274554 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90124024-daf7-4241-a787-2840359fe074-dns-svc\") pod \"dnsmasq-dns-6b7b667979-qdtmw\" (UID: \"90124024-daf7-4241-a787-2840359fe074\") " pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.275351 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/90124024-daf7-4241-a787-2840359fe074-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-qdtmw\" (UID: \"90124024-daf7-4241-a787-2840359fe074\") " pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.275383 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90124024-daf7-4241-a787-2840359fe074-config\") pod \"dnsmasq-dns-6b7b667979-qdtmw\" (UID: \"90124024-daf7-4241-a787-2840359fe074\") " pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.275502 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90124024-daf7-4241-a787-2840359fe074-dns-svc\") pod \"dnsmasq-dns-6b7b667979-qdtmw\" (UID: \"90124024-daf7-4241-a787-2840359fe074\") " pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.275912 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/90124024-daf7-4241-a787-2840359fe074-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-qdtmw\" (UID: \"90124024-daf7-4241-a787-2840359fe074\") " pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.276012 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/90124024-daf7-4241-a787-2840359fe074-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-qdtmw\" (UID: \"90124024-daf7-4241-a787-2840359fe074\") " pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.295574 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c79d794d7-tbd65" podUID="274211c2-377d-4164-be60-ca42e8167849" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: i/o timeout" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.315132 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2tb4\" (UniqueName: \"kubernetes.io/projected/90124024-daf7-4241-a787-2840359fe074-kube-api-access-c2tb4\") pod \"dnsmasq-dns-6b7b667979-qdtmw\" (UID: \"90124024-daf7-4241-a787-2840359fe074\") " pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.346086 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.376580 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c583806-afc7-4f95-8d4c-f0690157ca65-ovndb-tls-certs\") pod \"neutron-7c8fc98b5b-dwkdc\" (UID: \"7c583806-afc7-4f95-8d4c-f0690157ca65\") " pod="openstack/neutron-7c8fc98b5b-dwkdc" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.376630 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42rxx\" (UniqueName: \"kubernetes.io/projected/7c583806-afc7-4f95-8d4c-f0690157ca65-kube-api-access-42rxx\") pod \"neutron-7c8fc98b5b-dwkdc\" (UID: \"7c583806-afc7-4f95-8d4c-f0690157ca65\") " pod="openstack/neutron-7c8fc98b5b-dwkdc" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.376779 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7c583806-afc7-4f95-8d4c-f0690157ca65-httpd-config\") pod \"neutron-7c8fc98b5b-dwkdc\" (UID: \"7c583806-afc7-4f95-8d4c-f0690157ca65\") " pod="openstack/neutron-7c8fc98b5b-dwkdc" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.376822 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7c583806-afc7-4f95-8d4c-f0690157ca65-config\") pod \"neutron-7c8fc98b5b-dwkdc\" (UID: \"7c583806-afc7-4f95-8d4c-f0690157ca65\") " pod="openstack/neutron-7c8fc98b5b-dwkdc" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.376991 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c583806-afc7-4f95-8d4c-f0690157ca65-combined-ca-bundle\") pod \"neutron-7c8fc98b5b-dwkdc\" (UID: \"7c583806-afc7-4f95-8d4c-f0690157ca65\") " pod="openstack/neutron-7c8fc98b5b-dwkdc" Dec 06 16:03:15 crc kubenswrapper[4813]: W1206 16:03:15.440398 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28238998_5d44_4ae9_8d1f_93d56ff18152.slice/crio-144b936f94068463fae51e84a4ad48555d86367584965d2a0d6a24a8a934fa30 WatchSource:0}: Error finding container 144b936f94068463fae51e84a4ad48555d86367584965d2a0d6a24a8a934fa30: Status 404 returned error can't find the container with id 144b936f94068463fae51e84a4ad48555d86367584965d2a0d6a24a8a934fa30 Dec 06 16:03:15 crc kubenswrapper[4813]: E1206 16:03:15.443060 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 06 16:03:15 crc kubenswrapper[4813]: E1206 16:03:15.443166 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n679h94h5d9h56fh598h57fh5b6h5d9h65ch646hffh64h5b9h57dhfdh646h658h548h74h7bhd4h5b4h67ch68h675h596h565h5b9h57fh646h565h684q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bsqxj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-86c95686f7-d9tjn_openstack(4845d02f-a80d-4954-bede-716091ba57b8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 16:03:15 crc kubenswrapper[4813]: E1206 16:03:15.445006 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-86c95686f7-d9tjn" podUID="4845d02f-a80d-4954-bede-716091ba57b8" Dec 06 16:03:15 crc kubenswrapper[4813]: E1206 16:03:15.451467 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 06 16:03:15 crc kubenswrapper[4813]: E1206 16:03:15.451622 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cpz7c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-lvw7n_openstack(8cf4a1cf-ef96-4e50-946c-cd0c5eccff32): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 16:03:15 crc kubenswrapper[4813]: E1206 16:03:15.452760 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-lvw7n" podUID="8cf4a1cf-ef96-4e50-946c-cd0c5eccff32" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.479188 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c583806-afc7-4f95-8d4c-f0690157ca65-ovndb-tls-certs\") pod \"neutron-7c8fc98b5b-dwkdc\" (UID: \"7c583806-afc7-4f95-8d4c-f0690157ca65\") " pod="openstack/neutron-7c8fc98b5b-dwkdc" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.479235 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42rxx\" (UniqueName: \"kubernetes.io/projected/7c583806-afc7-4f95-8d4c-f0690157ca65-kube-api-access-42rxx\") pod \"neutron-7c8fc98b5b-dwkdc\" (UID: \"7c583806-afc7-4f95-8d4c-f0690157ca65\") " pod="openstack/neutron-7c8fc98b5b-dwkdc" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.479284 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7c583806-afc7-4f95-8d4c-f0690157ca65-httpd-config\") pod \"neutron-7c8fc98b5b-dwkdc\" (UID: \"7c583806-afc7-4f95-8d4c-f0690157ca65\") " pod="openstack/neutron-7c8fc98b5b-dwkdc" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.479325 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7c583806-afc7-4f95-8d4c-f0690157ca65-config\") pod \"neutron-7c8fc98b5b-dwkdc\" (UID: \"7c583806-afc7-4f95-8d4c-f0690157ca65\") " pod="openstack/neutron-7c8fc98b5b-dwkdc" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.479532 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c583806-afc7-4f95-8d4c-f0690157ca65-combined-ca-bundle\") pod \"neutron-7c8fc98b5b-dwkdc\" (UID: \"7c583806-afc7-4f95-8d4c-f0690157ca65\") " pod="openstack/neutron-7c8fc98b5b-dwkdc" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.497672 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c583806-afc7-4f95-8d4c-f0690157ca65-ovndb-tls-certs\") pod \"neutron-7c8fc98b5b-dwkdc\" (UID: \"7c583806-afc7-4f95-8d4c-f0690157ca65\") " pod="openstack/neutron-7c8fc98b5b-dwkdc" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.497767 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7c583806-afc7-4f95-8d4c-f0690157ca65-httpd-config\") pod \"neutron-7c8fc98b5b-dwkdc\" (UID: \"7c583806-afc7-4f95-8d4c-f0690157ca65\") " pod="openstack/neutron-7c8fc98b5b-dwkdc" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.498371 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/7c583806-afc7-4f95-8d4c-f0690157ca65-config\") pod \"neutron-7c8fc98b5b-dwkdc\" (UID: \"7c583806-afc7-4f95-8d4c-f0690157ca65\") " pod="openstack/neutron-7c8fc98b5b-dwkdc" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.517006 4813 scope.go:117] "RemoveContainer" containerID="3f280ff8667572d95419dbd6b782b0534fca40eb5d29d4cd3b6d0457f3a49c36" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.517839 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c583806-afc7-4f95-8d4c-f0690157ca65-combined-ca-bundle\") pod \"neutron-7c8fc98b5b-dwkdc\" (UID: \"7c583806-afc7-4f95-8d4c-f0690157ca65\") " pod="openstack/neutron-7c8fc98b5b-dwkdc" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.520956 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42rxx\" (UniqueName: \"kubernetes.io/projected/7c583806-afc7-4f95-8d4c-f0690157ca65-kube-api-access-42rxx\") pod \"neutron-7c8fc98b5b-dwkdc\" (UID: \"7c583806-afc7-4f95-8d4c-f0690157ca65\") " pod="openstack/neutron-7c8fc98b5b-dwkdc" Dec 06 16:03:15 crc kubenswrapper[4813]: I1206 16:03:15.545957 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7c8fc98b5b-dwkdc" Dec 06 16:03:16 crc kubenswrapper[4813]: I1206 16:03:16.029120 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-bbc668d58-h8w67"] Dec 06 16:03:16 crc kubenswrapper[4813]: I1206 16:03:16.259362 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-6qjfq"] Dec 06 16:03:16 crc kubenswrapper[4813]: I1206 16:03:16.316239 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-bbc668d58-h8w67" event={"ID":"4bdc8329-162b-4592-b98e-ec7eda7f2ce1","Type":"ContainerStarted","Data":"eb41bb3b1c841fb4eb3841ac4832b4adeefc0eb6a531069ccd6879a1ba4e6ef9"} Dec 06 16:03:16 crc kubenswrapper[4813]: I1206 16:03:16.322880 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fd6f7946b-kmm96" event={"ID":"28238998-5d44-4ae9-8d1f-93d56ff18152","Type":"ContainerStarted","Data":"144b936f94068463fae51e84a4ad48555d86367584965d2a0d6a24a8a934fa30"} Dec 06 16:03:16 crc kubenswrapper[4813]: E1206 16:03:16.385520 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-lvw7n" podUID="8cf4a1cf-ef96-4e50-946c-cd0c5eccff32" Dec 06 16:03:16 crc kubenswrapper[4813]: I1206 16:03:16.388478 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 06 16:03:16 crc kubenswrapper[4813]: I1206 16:03:16.817958 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7c8fc98b5b-dwkdc"] Dec 06 16:03:16 crc kubenswrapper[4813]: I1206 16:03:16.843746 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-qdtmw"] Dec 06 16:03:16 crc kubenswrapper[4813]: W1206 16:03:16.869395 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod90124024_daf7_4241_a787_2840359fe074.slice/crio-59262c1305d593a6033d2261739d33f8ab5c31d54f9b90e00777e37328674810 WatchSource:0}: Error finding container 59262c1305d593a6033d2261739d33f8ab5c31d54f9b90e00777e37328674810: Status 404 returned error can't find the container with id 59262c1305d593a6033d2261739d33f8ab5c31d54f9b90e00777e37328674810 Dec 06 16:03:16 crc kubenswrapper[4813]: I1206 16:03:16.962780 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-86c95686f7-d9tjn" Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.030615 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4845d02f-a80d-4954-bede-716091ba57b8-config-data\") pod \"4845d02f-a80d-4954-bede-716091ba57b8\" (UID: \"4845d02f-a80d-4954-bede-716091ba57b8\") " Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.030680 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4845d02f-a80d-4954-bede-716091ba57b8-horizon-secret-key\") pod \"4845d02f-a80d-4954-bede-716091ba57b8\" (UID: \"4845d02f-a80d-4954-bede-716091ba57b8\") " Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.030885 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4845d02f-a80d-4954-bede-716091ba57b8-logs\") pod \"4845d02f-a80d-4954-bede-716091ba57b8\" (UID: \"4845d02f-a80d-4954-bede-716091ba57b8\") " Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.030931 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4845d02f-a80d-4954-bede-716091ba57b8-scripts\") pod \"4845d02f-a80d-4954-bede-716091ba57b8\" (UID: \"4845d02f-a80d-4954-bede-716091ba57b8\") " Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.030954 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bsqxj\" (UniqueName: \"kubernetes.io/projected/4845d02f-a80d-4954-bede-716091ba57b8-kube-api-access-bsqxj\") pod \"4845d02f-a80d-4954-bede-716091ba57b8\" (UID: \"4845d02f-a80d-4954-bede-716091ba57b8\") " Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.031114 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4845d02f-a80d-4954-bede-716091ba57b8-config-data" (OuterVolumeSpecName: "config-data") pod "4845d02f-a80d-4954-bede-716091ba57b8" (UID: "4845d02f-a80d-4954-bede-716091ba57b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.031597 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4845d02f-a80d-4954-bede-716091ba57b8-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.032645 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4845d02f-a80d-4954-bede-716091ba57b8-logs" (OuterVolumeSpecName: "logs") pod "4845d02f-a80d-4954-bede-716091ba57b8" (UID: "4845d02f-a80d-4954-bede-716091ba57b8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.033379 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4845d02f-a80d-4954-bede-716091ba57b8-scripts" (OuterVolumeSpecName: "scripts") pod "4845d02f-a80d-4954-bede-716091ba57b8" (UID: "4845d02f-a80d-4954-bede-716091ba57b8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.038713 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4845d02f-a80d-4954-bede-716091ba57b8-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "4845d02f-a80d-4954-bede-716091ba57b8" (UID: "4845d02f-a80d-4954-bede-716091ba57b8"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.038842 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4845d02f-a80d-4954-bede-716091ba57b8-kube-api-access-bsqxj" (OuterVolumeSpecName: "kube-api-access-bsqxj") pod "4845d02f-a80d-4954-bede-716091ba57b8" (UID: "4845d02f-a80d-4954-bede-716091ba57b8"). InnerVolumeSpecName "kube-api-access-bsqxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.133382 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4845d02f-a80d-4954-bede-716091ba57b8-logs\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.133408 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4845d02f-a80d-4954-bede-716091ba57b8-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.133418 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bsqxj\" (UniqueName: \"kubernetes.io/projected/4845d02f-a80d-4954-bede-716091ba57b8-kube-api-access-bsqxj\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.133429 4813 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4845d02f-a80d-4954-bede-716091ba57b8-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.368312 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"372958b7-ef0c-448e-8acc-57a1b27a985f","Type":"ContainerStarted","Data":"355f363d43003c5cb046b813caf8e038695dd83c0745944264ddf9533760e167"} Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.368415 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="372958b7-ef0c-448e-8acc-57a1b27a985f" containerName="glance-log" containerID="cri-o://16b64922c56fc8c8c92c7d2d1584e849542a076ea94fe66abe7f0f797c9d706a" gracePeriod=30 Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.368681 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="372958b7-ef0c-448e-8acc-57a1b27a985f" containerName="glance-httpd" containerID="cri-o://355f363d43003c5cb046b813caf8e038695dd83c0745944264ddf9533760e167" gracePeriod=30 Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.382574 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7c8fc98b5b-dwkdc" event={"ID":"7c583806-afc7-4f95-8d4c-f0690157ca65","Type":"ContainerStarted","Data":"9999291e2b8d7565b054e93fa369aa0a9392136209dae8df114b56c191ed251a"} Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.382614 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7c8fc98b5b-dwkdc" event={"ID":"7c583806-afc7-4f95-8d4c-f0690157ca65","Type":"ContainerStarted","Data":"877e6f39eec698c0411af6e3176db9a7abb1924e940b11e11f1627535bf4dd13"} Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.393157 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nrdxf" event={"ID":"0179e5b2-8549-4810-83ab-0acbff2a5c6c","Type":"ContainerStarted","Data":"31c053cc89a499a51c9bf185c2591dc385e1869e2770b4367925745f679c2804"} Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.411190 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fd6f7946b-kmm96" event={"ID":"28238998-5d44-4ae9-8d1f-93d56ff18152","Type":"ContainerStarted","Data":"a345437efaa90d8997d33602751c98e878b581f8dff95a316322283821c67c59"} Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.411234 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fd6f7946b-kmm96" event={"ID":"28238998-5d44-4ae9-8d1f-93d56ff18152","Type":"ContainerStarted","Data":"cb15155904431296fb0be57bd7b6ec1bfb0599ff5a3f9edf808b80f9d842e3f9"} Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.415245 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-86c95686f7-d9tjn" Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.421433 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-86c95686f7-d9tjn" event={"ID":"4845d02f-a80d-4954-bede-716091ba57b8","Type":"ContainerDied","Data":"3bb2356ba7e868fb83f42a0a4ca8718b964a8862f2b0bba7261330de4fcf42c0"} Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.437128 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-bbc668d58-h8w67" event={"ID":"4bdc8329-162b-4592-b98e-ec7eda7f2ce1","Type":"ContainerStarted","Data":"3307007b53464dbc4d2a32eb29a1577325f2d8df6457bc6adf713e100058d865"} Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.437174 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-bbc668d58-h8w67" event={"ID":"4bdc8329-162b-4592-b98e-ec7eda7f2ce1","Type":"ContainerStarted","Data":"859333a385a1f24c723ad6a5689eae7998a7664c905e033301fe7108cfa531de"} Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.437583 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=40.437567634 podStartE2EDuration="40.437567634s" podCreationTimestamp="2025-12-06 16:02:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:03:17.435012997 +0000 UTC m=+1037.325892573" watchObservedRunningTime="2025-12-06 16:03:17.437567634 +0000 UTC m=+1037.328447210" Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.443725 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-qk8sl" event={"ID":"5c1f8568-2368-445f-a2cb-01ab3a6d1564","Type":"ContainerStarted","Data":"6033edd6c54603f70163c37784357d9aa1b8fa35bfeb8579b6476ddaaabc69cf"} Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.453392 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6qjfq" event={"ID":"67cf9ebe-1a58-4e50-9c10-f6f7933e781f","Type":"ContainerStarted","Data":"0a3cd73bf8fefc94389973058f01beb8ba5cf10cdc7fb883c3651d2a34f9ed30"} Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.453432 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6qjfq" event={"ID":"67cf9ebe-1a58-4e50-9c10-f6f7933e781f","Type":"ContainerStarted","Data":"4b25e79d792b968cd8c068aafbe3d9c04158ce7f04d3054381534631e83b5490"} Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.457978 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c4340aaa-8da3-4293-87f2-d49f43ee01bd","Type":"ContainerStarted","Data":"029171a9d8d467df307af88ca83fb4ba5a86af94672060aa466be0f458fa210e"} Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.458134 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="c4340aaa-8da3-4293-87f2-d49f43ee01bd" containerName="glance-log" containerID="cri-o://ddb27420ae62899fa3986c3835f7f9f44fb5d00e9f89858ac6a7ec912af9a51f" gracePeriod=30 Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.458454 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="c4340aaa-8da3-4293-87f2-d49f43ee01bd" containerName="glance-httpd" containerID="cri-o://029171a9d8d467df307af88ca83fb4ba5a86af94672060aa466be0f458fa210e" gracePeriod=30 Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.465298 4813 generic.go:334] "Generic (PLEG): container finished" podID="90124024-daf7-4241-a787-2840359fe074" containerID="0a7a2ceb51aff58c096e8545422f4025de69e0d87601130d3445d785e5a918ce" exitCode=0 Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.465391 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" event={"ID":"90124024-daf7-4241-a787-2840359fe074","Type":"ContainerDied","Data":"0a7a2ceb51aff58c096e8545422f4025de69e0d87601130d3445d785e5a918ce"} Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.465467 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" event={"ID":"90124024-daf7-4241-a787-2840359fe074","Type":"ContainerStarted","Data":"59262c1305d593a6033d2261739d33f8ab5c31d54f9b90e00777e37328674810"} Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.476281 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7fd6f7946b-kmm96" podStartSLOduration=31.905000978 podStartE2EDuration="32.476249098s" podCreationTimestamp="2025-12-06 16:02:45 +0000 UTC" firstStartedPulling="2025-12-06 16:03:15.443216686 +0000 UTC m=+1035.334096262" lastFinishedPulling="2025-12-06 16:03:16.014464806 +0000 UTC m=+1035.905344382" observedRunningTime="2025-12-06 16:03:17.46232606 +0000 UTC m=+1037.353205626" watchObservedRunningTime="2025-12-06 16:03:17.476249098 +0000 UTC m=+1037.367128674" Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.506755 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-nrdxf" podStartSLOduration=4.1641432720000005 podStartE2EDuration="40.506720755s" podCreationTimestamp="2025-12-06 16:02:37 +0000 UTC" firstStartedPulling="2025-12-06 16:02:39.118868895 +0000 UTC m=+999.009748471" lastFinishedPulling="2025-12-06 16:03:15.461446378 +0000 UTC m=+1035.352325954" observedRunningTime="2025-12-06 16:03:17.495877078 +0000 UTC m=+1037.386756644" watchObservedRunningTime="2025-12-06 16:03:17.506720755 +0000 UTC m=+1037.397600331" Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.545335 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-qk8sl" podStartSLOduration=3.849260047 podStartE2EDuration="40.545320066s" podCreationTimestamp="2025-12-06 16:02:37 +0000 UTC" firstStartedPulling="2025-12-06 16:02:38.744399734 +0000 UTC m=+998.635279310" lastFinishedPulling="2025-12-06 16:03:15.440459753 +0000 UTC m=+1035.331339329" observedRunningTime="2025-12-06 16:03:17.540319974 +0000 UTC m=+1037.431199550" watchObservedRunningTime="2025-12-06 16:03:17.545320066 +0000 UTC m=+1037.436199642" Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.625422 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-bbc668d58-h8w67" podStartSLOduration=32.625400806 podStartE2EDuration="32.625400806s" podCreationTimestamp="2025-12-06 16:02:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:03:17.56510281 +0000 UTC m=+1037.455982386" watchObservedRunningTime="2025-12-06 16:03:17.625400806 +0000 UTC m=+1037.516280382" Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.734297 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=40.734246617 podStartE2EDuration="40.734246617s" podCreationTimestamp="2025-12-06 16:02:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:03:17.588803247 +0000 UTC m=+1037.479682823" watchObservedRunningTime="2025-12-06 16:03:17.734246617 +0000 UTC m=+1037.625126193" Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.754249 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-6qjfq" podStartSLOduration=26.754224076 podStartE2EDuration="26.754224076s" podCreationTimestamp="2025-12-06 16:02:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:03:17.612154925 +0000 UTC m=+1037.503034491" watchObservedRunningTime="2025-12-06 16:03:17.754224076 +0000 UTC m=+1037.645103642" Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.796552 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-86c95686f7-d9tjn"] Dec 06 16:03:17 crc kubenswrapper[4813]: I1206 16:03:17.802823 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-86c95686f7-d9tjn"] Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.469801 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-759b8b45fc-psf25"] Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.471120 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-759b8b45fc-psf25" Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.480609 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.480619 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.491474 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-759b8b45fc-psf25"] Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.531424 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4845d02f-a80d-4954-bede-716091ba57b8" path="/var/lib/kubelet/pods/4845d02f-a80d-4954-bede-716091ba57b8/volumes" Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.536747 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7c8fc98b5b-dwkdc" event={"ID":"7c583806-afc7-4f95-8d4c-f0690157ca65","Type":"ContainerStarted","Data":"d6326e7e0b145e3ecd7b906844f31d37d84c0c433cced12074a74eecb1eb6d0d"} Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.536922 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7c8fc98b5b-dwkdc" Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.565910 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/afd03411-dfea-469b-804c-7cbe8a964d54-public-tls-certs\") pod \"neutron-759b8b45fc-psf25\" (UID: \"afd03411-dfea-469b-804c-7cbe8a964d54\") " pod="openstack/neutron-759b8b45fc-psf25" Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.565986 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/afd03411-dfea-469b-804c-7cbe8a964d54-config\") pod \"neutron-759b8b45fc-psf25\" (UID: \"afd03411-dfea-469b-804c-7cbe8a964d54\") " pod="openstack/neutron-759b8b45fc-psf25" Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.566014 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/afd03411-dfea-469b-804c-7cbe8a964d54-ovndb-tls-certs\") pod \"neutron-759b8b45fc-psf25\" (UID: \"afd03411-dfea-469b-804c-7cbe8a964d54\") " pod="openstack/neutron-759b8b45fc-psf25" Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.566032 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rfhh\" (UniqueName: \"kubernetes.io/projected/afd03411-dfea-469b-804c-7cbe8a964d54-kube-api-access-9rfhh\") pod \"neutron-759b8b45fc-psf25\" (UID: \"afd03411-dfea-469b-804c-7cbe8a964d54\") " pod="openstack/neutron-759b8b45fc-psf25" Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.566054 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/afd03411-dfea-469b-804c-7cbe8a964d54-httpd-config\") pod \"neutron-759b8b45fc-psf25\" (UID: \"afd03411-dfea-469b-804c-7cbe8a964d54\") " pod="openstack/neutron-759b8b45fc-psf25" Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.566101 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afd03411-dfea-469b-804c-7cbe8a964d54-combined-ca-bundle\") pod \"neutron-759b8b45fc-psf25\" (UID: \"afd03411-dfea-469b-804c-7cbe8a964d54\") " pod="openstack/neutron-759b8b45fc-psf25" Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.566118 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/afd03411-dfea-469b-804c-7cbe8a964d54-internal-tls-certs\") pod \"neutron-759b8b45fc-psf25\" (UID: \"afd03411-dfea-469b-804c-7cbe8a964d54\") " pod="openstack/neutron-759b8b45fc-psf25" Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.570370 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7c8fc98b5b-dwkdc" podStartSLOduration=3.570357757 podStartE2EDuration="3.570357757s" podCreationTimestamp="2025-12-06 16:03:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:03:18.565419676 +0000 UTC m=+1038.456299252" watchObservedRunningTime="2025-12-06 16:03:18.570357757 +0000 UTC m=+1038.461237333" Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.571953 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" event={"ID":"90124024-daf7-4241-a787-2840359fe074","Type":"ContainerStarted","Data":"b4833ba9d04cff0f35659eeb3b203e748fe5c3214b9fed117882cb9e9972b2b8"} Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.572694 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.578662 4813 generic.go:334] "Generic (PLEG): container finished" podID="c4340aaa-8da3-4293-87f2-d49f43ee01bd" containerID="029171a9d8d467df307af88ca83fb4ba5a86af94672060aa466be0f458fa210e" exitCode=0 Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.578687 4813 generic.go:334] "Generic (PLEG): container finished" podID="c4340aaa-8da3-4293-87f2-d49f43ee01bd" containerID="ddb27420ae62899fa3986c3835f7f9f44fb5d00e9f89858ac6a7ec912af9a51f" exitCode=143 Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.578723 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c4340aaa-8da3-4293-87f2-d49f43ee01bd","Type":"ContainerDied","Data":"029171a9d8d467df307af88ca83fb4ba5a86af94672060aa466be0f458fa210e"} Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.578745 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c4340aaa-8da3-4293-87f2-d49f43ee01bd","Type":"ContainerDied","Data":"ddb27420ae62899fa3986c3835f7f9f44fb5d00e9f89858ac6a7ec912af9a51f"} Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.580210 4813 generic.go:334] "Generic (PLEG): container finished" podID="372958b7-ef0c-448e-8acc-57a1b27a985f" containerID="16b64922c56fc8c8c92c7d2d1584e849542a076ea94fe66abe7f0f797c9d706a" exitCode=143 Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.580464 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"372958b7-ef0c-448e-8acc-57a1b27a985f","Type":"ContainerDied","Data":"16b64922c56fc8c8c92c7d2d1584e849542a076ea94fe66abe7f0f797c9d706a"} Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.600859 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" podStartSLOduration=3.600842274 podStartE2EDuration="3.600842274s" podCreationTimestamp="2025-12-06 16:03:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:03:18.590543931 +0000 UTC m=+1038.481423507" watchObservedRunningTime="2025-12-06 16:03:18.600842274 +0000 UTC m=+1038.491721850" Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.668785 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/afd03411-dfea-469b-804c-7cbe8a964d54-public-tls-certs\") pod \"neutron-759b8b45fc-psf25\" (UID: \"afd03411-dfea-469b-804c-7cbe8a964d54\") " pod="openstack/neutron-759b8b45fc-psf25" Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.669119 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/afd03411-dfea-469b-804c-7cbe8a964d54-config\") pod \"neutron-759b8b45fc-psf25\" (UID: \"afd03411-dfea-469b-804c-7cbe8a964d54\") " pod="openstack/neutron-759b8b45fc-psf25" Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.669188 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/afd03411-dfea-469b-804c-7cbe8a964d54-ovndb-tls-certs\") pod \"neutron-759b8b45fc-psf25\" (UID: \"afd03411-dfea-469b-804c-7cbe8a964d54\") " pod="openstack/neutron-759b8b45fc-psf25" Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.669222 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rfhh\" (UniqueName: \"kubernetes.io/projected/afd03411-dfea-469b-804c-7cbe8a964d54-kube-api-access-9rfhh\") pod \"neutron-759b8b45fc-psf25\" (UID: \"afd03411-dfea-469b-804c-7cbe8a964d54\") " pod="openstack/neutron-759b8b45fc-psf25" Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.669287 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/afd03411-dfea-469b-804c-7cbe8a964d54-httpd-config\") pod \"neutron-759b8b45fc-psf25\" (UID: \"afd03411-dfea-469b-804c-7cbe8a964d54\") " pod="openstack/neutron-759b8b45fc-psf25" Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.669418 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afd03411-dfea-469b-804c-7cbe8a964d54-combined-ca-bundle\") pod \"neutron-759b8b45fc-psf25\" (UID: \"afd03411-dfea-469b-804c-7cbe8a964d54\") " pod="openstack/neutron-759b8b45fc-psf25" Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.669436 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/afd03411-dfea-469b-804c-7cbe8a964d54-internal-tls-certs\") pod \"neutron-759b8b45fc-psf25\" (UID: \"afd03411-dfea-469b-804c-7cbe8a964d54\") " pod="openstack/neutron-759b8b45fc-psf25" Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.676653 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afd03411-dfea-469b-804c-7cbe8a964d54-combined-ca-bundle\") pod \"neutron-759b8b45fc-psf25\" (UID: \"afd03411-dfea-469b-804c-7cbe8a964d54\") " pod="openstack/neutron-759b8b45fc-psf25" Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.681473 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/afd03411-dfea-469b-804c-7cbe8a964d54-public-tls-certs\") pod \"neutron-759b8b45fc-psf25\" (UID: \"afd03411-dfea-469b-804c-7cbe8a964d54\") " pod="openstack/neutron-759b8b45fc-psf25" Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.682747 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/afd03411-dfea-469b-804c-7cbe8a964d54-httpd-config\") pod \"neutron-759b8b45fc-psf25\" (UID: \"afd03411-dfea-469b-804c-7cbe8a964d54\") " pod="openstack/neutron-759b8b45fc-psf25" Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.683515 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/afd03411-dfea-469b-804c-7cbe8a964d54-config\") pod \"neutron-759b8b45fc-psf25\" (UID: \"afd03411-dfea-469b-804c-7cbe8a964d54\") " pod="openstack/neutron-759b8b45fc-psf25" Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.684359 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/afd03411-dfea-469b-804c-7cbe8a964d54-internal-tls-certs\") pod \"neutron-759b8b45fc-psf25\" (UID: \"afd03411-dfea-469b-804c-7cbe8a964d54\") " pod="openstack/neutron-759b8b45fc-psf25" Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.693720 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/afd03411-dfea-469b-804c-7cbe8a964d54-ovndb-tls-certs\") pod \"neutron-759b8b45fc-psf25\" (UID: \"afd03411-dfea-469b-804c-7cbe8a964d54\") " pod="openstack/neutron-759b8b45fc-psf25" Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.695958 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rfhh\" (UniqueName: \"kubernetes.io/projected/afd03411-dfea-469b-804c-7cbe8a964d54-kube-api-access-9rfhh\") pod \"neutron-759b8b45fc-psf25\" (UID: \"afd03411-dfea-469b-804c-7cbe8a964d54\") " pod="openstack/neutron-759b8b45fc-psf25" Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.872575 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-759b8b45fc-psf25" Dec 06 16:03:18 crc kubenswrapper[4813]: I1206 16:03:18.959611 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.075681 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4340aaa-8da3-4293-87f2-d49f43ee01bd-combined-ca-bundle\") pod \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.075727 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwq7q\" (UniqueName: \"kubernetes.io/projected/c4340aaa-8da3-4293-87f2-d49f43ee01bd-kube-api-access-vwq7q\") pod \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.075752 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4340aaa-8da3-4293-87f2-d49f43ee01bd-scripts\") pod \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.075773 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4340aaa-8da3-4293-87f2-d49f43ee01bd-config-data\") pod \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.075803 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.075836 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4340aaa-8da3-4293-87f2-d49f43ee01bd-internal-tls-certs\") pod \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.075908 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4340aaa-8da3-4293-87f2-d49f43ee01bd-logs\") pod \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.075944 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c4340aaa-8da3-4293-87f2-d49f43ee01bd-httpd-run\") pod \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\" (UID: \"c4340aaa-8da3-4293-87f2-d49f43ee01bd\") " Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.076659 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4340aaa-8da3-4293-87f2-d49f43ee01bd-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c4340aaa-8da3-4293-87f2-d49f43ee01bd" (UID: "c4340aaa-8da3-4293-87f2-d49f43ee01bd"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.081315 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4340aaa-8da3-4293-87f2-d49f43ee01bd-logs" (OuterVolumeSpecName: "logs") pod "c4340aaa-8da3-4293-87f2-d49f43ee01bd" (UID: "c4340aaa-8da3-4293-87f2-d49f43ee01bd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.083672 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "c4340aaa-8da3-4293-87f2-d49f43ee01bd" (UID: "c4340aaa-8da3-4293-87f2-d49f43ee01bd"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.093192 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4340aaa-8da3-4293-87f2-d49f43ee01bd-scripts" (OuterVolumeSpecName: "scripts") pod "c4340aaa-8da3-4293-87f2-d49f43ee01bd" (UID: "c4340aaa-8da3-4293-87f2-d49f43ee01bd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.097625 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4340aaa-8da3-4293-87f2-d49f43ee01bd-kube-api-access-vwq7q" (OuterVolumeSpecName: "kube-api-access-vwq7q") pod "c4340aaa-8da3-4293-87f2-d49f43ee01bd" (UID: "c4340aaa-8da3-4293-87f2-d49f43ee01bd"). InnerVolumeSpecName "kube-api-access-vwq7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.115584 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4340aaa-8da3-4293-87f2-d49f43ee01bd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c4340aaa-8da3-4293-87f2-d49f43ee01bd" (UID: "c4340aaa-8da3-4293-87f2-d49f43ee01bd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.172831 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4340aaa-8da3-4293-87f2-d49f43ee01bd-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c4340aaa-8da3-4293-87f2-d49f43ee01bd" (UID: "c4340aaa-8da3-4293-87f2-d49f43ee01bd"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.178130 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4340aaa-8da3-4293-87f2-d49f43ee01bd-logs\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.178166 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c4340aaa-8da3-4293-87f2-d49f43ee01bd-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.178176 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4340aaa-8da3-4293-87f2-d49f43ee01bd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.178186 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwq7q\" (UniqueName: \"kubernetes.io/projected/c4340aaa-8da3-4293-87f2-d49f43ee01bd-kube-api-access-vwq7q\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.178195 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4340aaa-8da3-4293-87f2-d49f43ee01bd-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.178222 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.178231 4813 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4340aaa-8da3-4293-87f2-d49f43ee01bd-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.182915 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4340aaa-8da3-4293-87f2-d49f43ee01bd-config-data" (OuterVolumeSpecName: "config-data") pod "c4340aaa-8da3-4293-87f2-d49f43ee01bd" (UID: "c4340aaa-8da3-4293-87f2-d49f43ee01bd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.210180 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.279636 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4340aaa-8da3-4293-87f2-d49f43ee01bd-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.279669 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.427294 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.427347 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.427390 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.428040 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"659cf1c13b1c009e3dc025bbc1b70aff2561005d40bd452ea21f924d8a1cca89"} pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.428087 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" containerID="cri-o://659cf1c13b1c009e3dc025bbc1b70aff2561005d40bd452ea21f924d8a1cca89" gracePeriod=600 Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.524501 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.585360 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/372958b7-ef0c-448e-8acc-57a1b27a985f-scripts\") pod \"372958b7-ef0c-448e-8acc-57a1b27a985f\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.585398 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/372958b7-ef0c-448e-8acc-57a1b27a985f-logs\") pod \"372958b7-ef0c-448e-8acc-57a1b27a985f\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.585471 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlz8h\" (UniqueName: \"kubernetes.io/projected/372958b7-ef0c-448e-8acc-57a1b27a985f-kube-api-access-rlz8h\") pod \"372958b7-ef0c-448e-8acc-57a1b27a985f\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.585496 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/372958b7-ef0c-448e-8acc-57a1b27a985f-public-tls-certs\") pod \"372958b7-ef0c-448e-8acc-57a1b27a985f\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.585528 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"372958b7-ef0c-448e-8acc-57a1b27a985f\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.585598 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/372958b7-ef0c-448e-8acc-57a1b27a985f-httpd-run\") pod \"372958b7-ef0c-448e-8acc-57a1b27a985f\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.585641 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/372958b7-ef0c-448e-8acc-57a1b27a985f-combined-ca-bundle\") pod \"372958b7-ef0c-448e-8acc-57a1b27a985f\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.585692 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/372958b7-ef0c-448e-8acc-57a1b27a985f-config-data\") pod \"372958b7-ef0c-448e-8acc-57a1b27a985f\" (UID: \"372958b7-ef0c-448e-8acc-57a1b27a985f\") " Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.585877 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/372958b7-ef0c-448e-8acc-57a1b27a985f-logs" (OuterVolumeSpecName: "logs") pod "372958b7-ef0c-448e-8acc-57a1b27a985f" (UID: "372958b7-ef0c-448e-8acc-57a1b27a985f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.586476 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/372958b7-ef0c-448e-8acc-57a1b27a985f-logs\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.589934 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/372958b7-ef0c-448e-8acc-57a1b27a985f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "372958b7-ef0c-448e-8acc-57a1b27a985f" (UID: "372958b7-ef0c-448e-8acc-57a1b27a985f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.590364 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "372958b7-ef0c-448e-8acc-57a1b27a985f" (UID: "372958b7-ef0c-448e-8acc-57a1b27a985f"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.606349 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/372958b7-ef0c-448e-8acc-57a1b27a985f-scripts" (OuterVolumeSpecName: "scripts") pod "372958b7-ef0c-448e-8acc-57a1b27a985f" (UID: "372958b7-ef0c-448e-8acc-57a1b27a985f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.613861 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/372958b7-ef0c-448e-8acc-57a1b27a985f-kube-api-access-rlz8h" (OuterVolumeSpecName: "kube-api-access-rlz8h") pod "372958b7-ef0c-448e-8acc-57a1b27a985f" (UID: "372958b7-ef0c-448e-8acc-57a1b27a985f"). InnerVolumeSpecName "kube-api-access-rlz8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.615952 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c4340aaa-8da3-4293-87f2-d49f43ee01bd","Type":"ContainerDied","Data":"4409c57b9e5fee22f6438ee07f0bd9b1da473524285dc380bfb6dd1c8c70defa"} Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.615996 4813 scope.go:117] "RemoveContainer" containerID="029171a9d8d467df307af88ca83fb4ba5a86af94672060aa466be0f458fa210e" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.616106 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.629638 4813 generic.go:334] "Generic (PLEG): container finished" podID="372958b7-ef0c-448e-8acc-57a1b27a985f" containerID="355f363d43003c5cb046b813caf8e038695dd83c0745944264ddf9533760e167" exitCode=0 Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.629843 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"372958b7-ef0c-448e-8acc-57a1b27a985f","Type":"ContainerDied","Data":"355f363d43003c5cb046b813caf8e038695dd83c0745944264ddf9533760e167"} Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.629996 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"372958b7-ef0c-448e-8acc-57a1b27a985f","Type":"ContainerDied","Data":"2c70f8c53d8879b8a98ba274d20289687daf28521627dd9e5025946c0082699c"} Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.630111 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.672553 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-759b8b45fc-psf25"] Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.689749 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/372958b7-ef0c-448e-8acc-57a1b27a985f-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.689833 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/372958b7-ef0c-448e-8acc-57a1b27a985f-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.689884 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlz8h\" (UniqueName: \"kubernetes.io/projected/372958b7-ef0c-448e-8acc-57a1b27a985f-kube-api-access-rlz8h\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.689962 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.706527 4813 generic.go:334] "Generic (PLEG): container finished" podID="d88e8bae-c055-4c55-b548-f621ff96de06" containerID="659cf1c13b1c009e3dc025bbc1b70aff2561005d40bd452ea21f924d8a1cca89" exitCode=0 Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.706628 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerDied","Data":"659cf1c13b1c009e3dc025bbc1b70aff2561005d40bd452ea21f924d8a1cca89"} Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.720941 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/372958b7-ef0c-448e-8acc-57a1b27a985f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "372958b7-ef0c-448e-8acc-57a1b27a985f" (UID: "372958b7-ef0c-448e-8acc-57a1b27a985f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.737455 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f539e8ab-300c-426b-98e7-f9e87a6787b5","Type":"ContainerStarted","Data":"dfbc16c809c9610b60841cb1e75b5a4aa13ade73ea1b1707f16b77a327c14418"} Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.753326 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.758140 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.761511 4813 scope.go:117] "RemoveContainer" containerID="ddb27420ae62899fa3986c3835f7f9f44fb5d00e9f89858ac6a7ec912af9a51f" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.772103 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.797339 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 16:03:19 crc kubenswrapper[4813]: E1206 16:03:19.810923 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="372958b7-ef0c-448e-8acc-57a1b27a985f" containerName="glance-log" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.821350 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="372958b7-ef0c-448e-8acc-57a1b27a985f" containerName="glance-log" Dec 06 16:03:19 crc kubenswrapper[4813]: E1206 16:03:19.821400 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="372958b7-ef0c-448e-8acc-57a1b27a985f" containerName="glance-httpd" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.821407 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="372958b7-ef0c-448e-8acc-57a1b27a985f" containerName="glance-httpd" Dec 06 16:03:19 crc kubenswrapper[4813]: E1206 16:03:19.821486 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4340aaa-8da3-4293-87f2-d49f43ee01bd" containerName="glance-httpd" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.821495 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4340aaa-8da3-4293-87f2-d49f43ee01bd" containerName="glance-httpd" Dec 06 16:03:19 crc kubenswrapper[4813]: E1206 16:03:19.821512 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4340aaa-8da3-4293-87f2-d49f43ee01bd" containerName="glance-log" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.821519 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4340aaa-8da3-4293-87f2-d49f43ee01bd" containerName="glance-log" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.838643 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="372958b7-ef0c-448e-8acc-57a1b27a985f" containerName="glance-log" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.838686 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4340aaa-8da3-4293-87f2-d49f43ee01bd" containerName="glance-log" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.838695 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="372958b7-ef0c-448e-8acc-57a1b27a985f" containerName="glance-httpd" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.838732 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4340aaa-8da3-4293-87f2-d49f43ee01bd" containerName="glance-httpd" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.840687 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.848467 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.848709 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.849213 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/372958b7-ef0c-448e-8acc-57a1b27a985f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.849237 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.870218 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.873756 4813 scope.go:117] "RemoveContainer" containerID="355f363d43003c5cb046b813caf8e038695dd83c0745944264ddf9533760e167" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.924897 4813 scope.go:117] "RemoveContainer" containerID="16b64922c56fc8c8c92c7d2d1584e849542a076ea94fe66abe7f0f797c9d706a" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.925443 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/372958b7-ef0c-448e-8acc-57a1b27a985f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "372958b7-ef0c-448e-8acc-57a1b27a985f" (UID: "372958b7-ef0c-448e-8acc-57a1b27a985f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.952599 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aa78964-46a9-4784-8d7a-9762bac90670-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.952689 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6aa78964-46a9-4784-8d7a-9762bac90670-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.952732 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6aa78964-46a9-4784-8d7a-9762bac90670-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.952756 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.952799 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6aa78964-46a9-4784-8d7a-9762bac90670-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.952862 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6aa78964-46a9-4784-8d7a-9762bac90670-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.952911 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-js27q\" (UniqueName: \"kubernetes.io/projected/6aa78964-46a9-4784-8d7a-9762bac90670-kube-api-access-js27q\") pod \"glance-default-internal-api-0\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.953917 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6aa78964-46a9-4784-8d7a-9762bac90670-logs\") pod \"glance-default-internal-api-0\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.954056 4813 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/372958b7-ef0c-448e-8acc-57a1b27a985f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.963777 4813 scope.go:117] "RemoveContainer" containerID="355f363d43003c5cb046b813caf8e038695dd83c0745944264ddf9533760e167" Dec 06 16:03:19 crc kubenswrapper[4813]: E1206 16:03:19.971250 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"355f363d43003c5cb046b813caf8e038695dd83c0745944264ddf9533760e167\": container with ID starting with 355f363d43003c5cb046b813caf8e038695dd83c0745944264ddf9533760e167 not found: ID does not exist" containerID="355f363d43003c5cb046b813caf8e038695dd83c0745944264ddf9533760e167" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.971304 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"355f363d43003c5cb046b813caf8e038695dd83c0745944264ddf9533760e167"} err="failed to get container status \"355f363d43003c5cb046b813caf8e038695dd83c0745944264ddf9533760e167\": rpc error: code = NotFound desc = could not find container \"355f363d43003c5cb046b813caf8e038695dd83c0745944264ddf9533760e167\": container with ID starting with 355f363d43003c5cb046b813caf8e038695dd83c0745944264ddf9533760e167 not found: ID does not exist" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.971328 4813 scope.go:117] "RemoveContainer" containerID="16b64922c56fc8c8c92c7d2d1584e849542a076ea94fe66abe7f0f797c9d706a" Dec 06 16:03:19 crc kubenswrapper[4813]: E1206 16:03:19.971756 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16b64922c56fc8c8c92c7d2d1584e849542a076ea94fe66abe7f0f797c9d706a\": container with ID starting with 16b64922c56fc8c8c92c7d2d1584e849542a076ea94fe66abe7f0f797c9d706a not found: ID does not exist" containerID="16b64922c56fc8c8c92c7d2d1584e849542a076ea94fe66abe7f0f797c9d706a" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.971776 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16b64922c56fc8c8c92c7d2d1584e849542a076ea94fe66abe7f0f797c9d706a"} err="failed to get container status \"16b64922c56fc8c8c92c7d2d1584e849542a076ea94fe66abe7f0f797c9d706a\": rpc error: code = NotFound desc = could not find container \"16b64922c56fc8c8c92c7d2d1584e849542a076ea94fe66abe7f0f797c9d706a\": container with ID starting with 16b64922c56fc8c8c92c7d2d1584e849542a076ea94fe66abe7f0f797c9d706a not found: ID does not exist" Dec 06 16:03:19 crc kubenswrapper[4813]: I1206 16:03:19.971797 4813 scope.go:117] "RemoveContainer" containerID="0d0cad4af55e6ecbb2a20d8a543c0848e24b8077a79dec0b4034636d4a17f514" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.001289 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/372958b7-ef0c-448e-8acc-57a1b27a985f-config-data" (OuterVolumeSpecName: "config-data") pod "372958b7-ef0c-448e-8acc-57a1b27a985f" (UID: "372958b7-ef0c-448e-8acc-57a1b27a985f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.055441 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6aa78964-46a9-4784-8d7a-9762bac90670-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.055514 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6aa78964-46a9-4784-8d7a-9762bac90670-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.055557 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-js27q\" (UniqueName: \"kubernetes.io/projected/6aa78964-46a9-4784-8d7a-9762bac90670-kube-api-access-js27q\") pod \"glance-default-internal-api-0\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.055579 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6aa78964-46a9-4784-8d7a-9762bac90670-logs\") pod \"glance-default-internal-api-0\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.055598 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aa78964-46a9-4784-8d7a-9762bac90670-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.055648 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6aa78964-46a9-4784-8d7a-9762bac90670-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.055672 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6aa78964-46a9-4784-8d7a-9762bac90670-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.055690 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.055729 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/372958b7-ef0c-448e-8acc-57a1b27a985f-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.055951 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.057700 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6aa78964-46a9-4784-8d7a-9762bac90670-logs\") pod \"glance-default-internal-api-0\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.060363 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6aa78964-46a9-4784-8d7a-9762bac90670-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.072960 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6aa78964-46a9-4784-8d7a-9762bac90670-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.073580 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aa78964-46a9-4784-8d7a-9762bac90670-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.073732 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6aa78964-46a9-4784-8d7a-9762bac90670-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.078608 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6aa78964-46a9-4784-8d7a-9762bac90670-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.082054 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-js27q\" (UniqueName: \"kubernetes.io/projected/6aa78964-46a9-4784-8d7a-9762bac90670-kube-api-access-js27q\") pod \"glance-default-internal-api-0\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.107269 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.191609 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.277498 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.310436 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.330005 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.331329 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.337236 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.347622 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.351750 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.462445 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lplm\" (UniqueName: \"kubernetes.io/projected/5bccad5f-06ef-46af-b5a7-f4b761b02550-kube-api-access-7lplm\") pod \"glance-default-external-api-0\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " pod="openstack/glance-default-external-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.462712 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bccad5f-06ef-46af-b5a7-f4b761b02550-config-data\") pod \"glance-default-external-api-0\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " pod="openstack/glance-default-external-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.462762 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bccad5f-06ef-46af-b5a7-f4b761b02550-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " pod="openstack/glance-default-external-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.462796 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bccad5f-06ef-46af-b5a7-f4b761b02550-logs\") pod \"glance-default-external-api-0\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " pod="openstack/glance-default-external-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.462831 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " pod="openstack/glance-default-external-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.462848 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bccad5f-06ef-46af-b5a7-f4b761b02550-scripts\") pod \"glance-default-external-api-0\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " pod="openstack/glance-default-external-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.462872 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5bccad5f-06ef-46af-b5a7-f4b761b02550-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " pod="openstack/glance-default-external-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.462894 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bccad5f-06ef-46af-b5a7-f4b761b02550-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " pod="openstack/glance-default-external-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.534668 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="372958b7-ef0c-448e-8acc-57a1b27a985f" path="/var/lib/kubelet/pods/372958b7-ef0c-448e-8acc-57a1b27a985f/volumes" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.535414 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4340aaa-8da3-4293-87f2-d49f43ee01bd" path="/var/lib/kubelet/pods/c4340aaa-8da3-4293-87f2-d49f43ee01bd/volumes" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.566477 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bccad5f-06ef-46af-b5a7-f4b761b02550-logs\") pod \"glance-default-external-api-0\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " pod="openstack/glance-default-external-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.566530 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " pod="openstack/glance-default-external-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.566584 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bccad5f-06ef-46af-b5a7-f4b761b02550-scripts\") pod \"glance-default-external-api-0\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " pod="openstack/glance-default-external-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.567026 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-external-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.567140 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5bccad5f-06ef-46af-b5a7-f4b761b02550-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " pod="openstack/glance-default-external-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.567171 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bccad5f-06ef-46af-b5a7-f4b761b02550-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " pod="openstack/glance-default-external-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.567245 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lplm\" (UniqueName: \"kubernetes.io/projected/5bccad5f-06ef-46af-b5a7-f4b761b02550-kube-api-access-7lplm\") pod \"glance-default-external-api-0\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " pod="openstack/glance-default-external-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.567310 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bccad5f-06ef-46af-b5a7-f4b761b02550-config-data\") pod \"glance-default-external-api-0\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " pod="openstack/glance-default-external-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.567352 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bccad5f-06ef-46af-b5a7-f4b761b02550-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " pod="openstack/glance-default-external-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.567458 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5bccad5f-06ef-46af-b5a7-f4b761b02550-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " pod="openstack/glance-default-external-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.567243 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bccad5f-06ef-46af-b5a7-f4b761b02550-logs\") pod \"glance-default-external-api-0\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " pod="openstack/glance-default-external-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.595167 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bccad5f-06ef-46af-b5a7-f4b761b02550-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " pod="openstack/glance-default-external-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.606468 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bccad5f-06ef-46af-b5a7-f4b761b02550-scripts\") pod \"glance-default-external-api-0\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " pod="openstack/glance-default-external-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.609003 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bccad5f-06ef-46af-b5a7-f4b761b02550-config-data\") pod \"glance-default-external-api-0\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " pod="openstack/glance-default-external-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.615864 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bccad5f-06ef-46af-b5a7-f4b761b02550-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " pod="openstack/glance-default-external-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.620755 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lplm\" (UniqueName: \"kubernetes.io/projected/5bccad5f-06ef-46af-b5a7-f4b761b02550-kube-api-access-7lplm\") pod \"glance-default-external-api-0\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " pod="openstack/glance-default-external-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.690399 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " pod="openstack/glance-default-external-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.787740 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerStarted","Data":"b7b68c16444a3326ba33662bc5c9583c0b689e2b31a8a7643be0512e87a18c96"} Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.792953 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-759b8b45fc-psf25" event={"ID":"afd03411-dfea-469b-804c-7cbe8a964d54","Type":"ContainerStarted","Data":"b3114fe3e085ae8e8387e246c0f94da72d60b2dfdbd89929b5a941855c52d113"} Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.793005 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-759b8b45fc-psf25" event={"ID":"afd03411-dfea-469b-804c-7cbe8a964d54","Type":"ContainerStarted","Data":"5325957edc1190ae6b5a3638737cb6779c06c924c935ac7bd42d041aaab11027"} Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.820649 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 16:03:20 crc kubenswrapper[4813]: I1206 16:03:20.926594 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 16:03:21 crc kubenswrapper[4813]: I1206 16:03:21.432728 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 16:03:21 crc kubenswrapper[4813]: I1206 16:03:21.815615 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-759b8b45fc-psf25" event={"ID":"afd03411-dfea-469b-804c-7cbe8a964d54","Type":"ContainerStarted","Data":"07ffb182d76cbfb1ed9a91a61e9dbe93633589a003b02d5555008983244f3af2"} Dec 06 16:03:21 crc kubenswrapper[4813]: I1206 16:03:21.816913 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-759b8b45fc-psf25" Dec 06 16:03:21 crc kubenswrapper[4813]: I1206 16:03:21.824007 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5bccad5f-06ef-46af-b5a7-f4b761b02550","Type":"ContainerStarted","Data":"29c2ce8c03fa93fedc67caadfcfaa11b0a06cb8e004ff0dd414d77465c54aa80"} Dec 06 16:03:21 crc kubenswrapper[4813]: I1206 16:03:21.825345 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6aa78964-46a9-4784-8d7a-9762bac90670","Type":"ContainerStarted","Data":"29adfd7844b15dbad3e74bde5271f288e44cf33a80d9e3c51e32eb734510cad5"} Dec 06 16:03:21 crc kubenswrapper[4813]: I1206 16:03:21.829940 4813 generic.go:334] "Generic (PLEG): container finished" podID="0179e5b2-8549-4810-83ab-0acbff2a5c6c" containerID="31c053cc89a499a51c9bf185c2591dc385e1869e2770b4367925745f679c2804" exitCode=0 Dec 06 16:03:21 crc kubenswrapper[4813]: I1206 16:03:21.830065 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nrdxf" event={"ID":"0179e5b2-8549-4810-83ab-0acbff2a5c6c","Type":"ContainerDied","Data":"31c053cc89a499a51c9bf185c2591dc385e1869e2770b4367925745f679c2804"} Dec 06 16:03:21 crc kubenswrapper[4813]: I1206 16:03:21.848445 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-759b8b45fc-psf25" podStartSLOduration=3.848428298 podStartE2EDuration="3.848428298s" podCreationTimestamp="2025-12-06 16:03:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:03:21.843365585 +0000 UTC m=+1041.734245161" watchObservedRunningTime="2025-12-06 16:03:21.848428298 +0000 UTC m=+1041.739307864" Dec 06 16:03:22 crc kubenswrapper[4813]: I1206 16:03:22.847390 4813 generic.go:334] "Generic (PLEG): container finished" podID="5c1f8568-2368-445f-a2cb-01ab3a6d1564" containerID="6033edd6c54603f70163c37784357d9aa1b8fa35bfeb8579b6476ddaaabc69cf" exitCode=0 Dec 06 16:03:22 crc kubenswrapper[4813]: I1206 16:03:22.847489 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-qk8sl" event={"ID":"5c1f8568-2368-445f-a2cb-01ab3a6d1564","Type":"ContainerDied","Data":"6033edd6c54603f70163c37784357d9aa1b8fa35bfeb8579b6476ddaaabc69cf"} Dec 06 16:03:22 crc kubenswrapper[4813]: I1206 16:03:22.851538 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5bccad5f-06ef-46af-b5a7-f4b761b02550","Type":"ContainerStarted","Data":"f7750f8c7a2adbdd23c031fe8057997104f67f39c52c41c224f654e6a65985e1"} Dec 06 16:03:22 crc kubenswrapper[4813]: I1206 16:03:22.855339 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6aa78964-46a9-4784-8d7a-9762bac90670","Type":"ContainerStarted","Data":"aac4a1741df703d35aaee5da55da2f600ff92c65b049582b59acb6099c789212"} Dec 06 16:03:23 crc kubenswrapper[4813]: I1206 16:03:23.399990 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nrdxf" Dec 06 16:03:23 crc kubenswrapper[4813]: I1206 16:03:23.533430 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0179e5b2-8549-4810-83ab-0acbff2a5c6c-scripts\") pod \"0179e5b2-8549-4810-83ab-0acbff2a5c6c\" (UID: \"0179e5b2-8549-4810-83ab-0acbff2a5c6c\") " Dec 06 16:03:23 crc kubenswrapper[4813]: I1206 16:03:23.533482 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0179e5b2-8549-4810-83ab-0acbff2a5c6c-config-data\") pod \"0179e5b2-8549-4810-83ab-0acbff2a5c6c\" (UID: \"0179e5b2-8549-4810-83ab-0acbff2a5c6c\") " Dec 06 16:03:23 crc kubenswrapper[4813]: I1206 16:03:23.533502 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0179e5b2-8549-4810-83ab-0acbff2a5c6c-combined-ca-bundle\") pod \"0179e5b2-8549-4810-83ab-0acbff2a5c6c\" (UID: \"0179e5b2-8549-4810-83ab-0acbff2a5c6c\") " Dec 06 16:03:23 crc kubenswrapper[4813]: I1206 16:03:23.533601 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzm7x\" (UniqueName: \"kubernetes.io/projected/0179e5b2-8549-4810-83ab-0acbff2a5c6c-kube-api-access-pzm7x\") pod \"0179e5b2-8549-4810-83ab-0acbff2a5c6c\" (UID: \"0179e5b2-8549-4810-83ab-0acbff2a5c6c\") " Dec 06 16:03:23 crc kubenswrapper[4813]: I1206 16:03:23.533676 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0179e5b2-8549-4810-83ab-0acbff2a5c6c-logs\") pod \"0179e5b2-8549-4810-83ab-0acbff2a5c6c\" (UID: \"0179e5b2-8549-4810-83ab-0acbff2a5c6c\") " Dec 06 16:03:23 crc kubenswrapper[4813]: I1206 16:03:23.534351 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0179e5b2-8549-4810-83ab-0acbff2a5c6c-logs" (OuterVolumeSpecName: "logs") pod "0179e5b2-8549-4810-83ab-0acbff2a5c6c" (UID: "0179e5b2-8549-4810-83ab-0acbff2a5c6c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:03:23 crc kubenswrapper[4813]: I1206 16:03:23.542309 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0179e5b2-8549-4810-83ab-0acbff2a5c6c-scripts" (OuterVolumeSpecName: "scripts") pod "0179e5b2-8549-4810-83ab-0acbff2a5c6c" (UID: "0179e5b2-8549-4810-83ab-0acbff2a5c6c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:23 crc kubenswrapper[4813]: I1206 16:03:23.557455 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0179e5b2-8549-4810-83ab-0acbff2a5c6c-kube-api-access-pzm7x" (OuterVolumeSpecName: "kube-api-access-pzm7x") pod "0179e5b2-8549-4810-83ab-0acbff2a5c6c" (UID: "0179e5b2-8549-4810-83ab-0acbff2a5c6c"). InnerVolumeSpecName "kube-api-access-pzm7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:03:23 crc kubenswrapper[4813]: I1206 16:03:23.577626 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0179e5b2-8549-4810-83ab-0acbff2a5c6c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0179e5b2-8549-4810-83ab-0acbff2a5c6c" (UID: "0179e5b2-8549-4810-83ab-0acbff2a5c6c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:23 crc kubenswrapper[4813]: I1206 16:03:23.591465 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0179e5b2-8549-4810-83ab-0acbff2a5c6c-config-data" (OuterVolumeSpecName: "config-data") pod "0179e5b2-8549-4810-83ab-0acbff2a5c6c" (UID: "0179e5b2-8549-4810-83ab-0acbff2a5c6c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:23 crc kubenswrapper[4813]: I1206 16:03:23.635277 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0179e5b2-8549-4810-83ab-0acbff2a5c6c-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:23 crc kubenswrapper[4813]: I1206 16:03:23.635312 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0179e5b2-8549-4810-83ab-0acbff2a5c6c-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:23 crc kubenswrapper[4813]: I1206 16:03:23.635324 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0179e5b2-8549-4810-83ab-0acbff2a5c6c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:23 crc kubenswrapper[4813]: I1206 16:03:23.635338 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzm7x\" (UniqueName: \"kubernetes.io/projected/0179e5b2-8549-4810-83ab-0acbff2a5c6c-kube-api-access-pzm7x\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:23 crc kubenswrapper[4813]: I1206 16:03:23.635352 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0179e5b2-8549-4810-83ab-0acbff2a5c6c-logs\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:23 crc kubenswrapper[4813]: I1206 16:03:23.868665 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5bccad5f-06ef-46af-b5a7-f4b761b02550","Type":"ContainerStarted","Data":"9af44f16657483a9f9bcb4d8995264621f9f361deb554a563ffaf3ecbe22f6d3"} Dec 06 16:03:23 crc kubenswrapper[4813]: I1206 16:03:23.876833 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6aa78964-46a9-4784-8d7a-9762bac90670","Type":"ContainerStarted","Data":"8d56b17b6ec7d2287eae2b2f27075a41e25619c58499a939e57419b069771a85"} Dec 06 16:03:23 crc kubenswrapper[4813]: I1206 16:03:23.880756 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nrdxf" event={"ID":"0179e5b2-8549-4810-83ab-0acbff2a5c6c","Type":"ContainerDied","Data":"7f9d76ce538710b4003662dfeefe3596f9f64e19669a9c29903182fd9bfb2d9a"} Dec 06 16:03:23 crc kubenswrapper[4813]: I1206 16:03:23.880799 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f9d76ce538710b4003662dfeefe3596f9f64e19669a9c29903182fd9bfb2d9a" Dec 06 16:03:23 crc kubenswrapper[4813]: I1206 16:03:23.880782 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nrdxf" Dec 06 16:03:23 crc kubenswrapper[4813]: I1206 16:03:23.917344 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.917327817 podStartE2EDuration="3.917327817s" podCreationTimestamp="2025-12-06 16:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:03:23.9022147 +0000 UTC m=+1043.793094276" watchObservedRunningTime="2025-12-06 16:03:23.917327817 +0000 UTC m=+1043.808207393" Dec 06 16:03:23 crc kubenswrapper[4813]: I1206 16:03:23.927460 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.927440683 podStartE2EDuration="4.927440683s" podCreationTimestamp="2025-12-06 16:03:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:03:23.926902189 +0000 UTC m=+1043.817781765" watchObservedRunningTime="2025-12-06 16:03:23.927440683 +0000 UTC m=+1043.818320259" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.054190 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-98dddfffb-l698f"] Dec 06 16:03:24 crc kubenswrapper[4813]: E1206 16:03:24.054850 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0179e5b2-8549-4810-83ab-0acbff2a5c6c" containerName="placement-db-sync" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.054868 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0179e5b2-8549-4810-83ab-0acbff2a5c6c" containerName="placement-db-sync" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.055106 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0179e5b2-8549-4810-83ab-0acbff2a5c6c" containerName="placement-db-sync" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.056236 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-98dddfffb-l698f" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.060527 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-kvctl" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.060658 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.060683 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.060591 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.060892 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.062879 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-98dddfffb-l698f"] Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.142400 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v96lx\" (UniqueName: \"kubernetes.io/projected/e913dfcd-dfe3-4c2c-b72a-8393295c95fc-kube-api-access-v96lx\") pod \"placement-98dddfffb-l698f\" (UID: \"e913dfcd-dfe3-4c2c-b72a-8393295c95fc\") " pod="openstack/placement-98dddfffb-l698f" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.142460 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e913dfcd-dfe3-4c2c-b72a-8393295c95fc-logs\") pod \"placement-98dddfffb-l698f\" (UID: \"e913dfcd-dfe3-4c2c-b72a-8393295c95fc\") " pod="openstack/placement-98dddfffb-l698f" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.142491 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e913dfcd-dfe3-4c2c-b72a-8393295c95fc-config-data\") pod \"placement-98dddfffb-l698f\" (UID: \"e913dfcd-dfe3-4c2c-b72a-8393295c95fc\") " pod="openstack/placement-98dddfffb-l698f" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.142520 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e913dfcd-dfe3-4c2c-b72a-8393295c95fc-public-tls-certs\") pod \"placement-98dddfffb-l698f\" (UID: \"e913dfcd-dfe3-4c2c-b72a-8393295c95fc\") " pod="openstack/placement-98dddfffb-l698f" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.142575 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e913dfcd-dfe3-4c2c-b72a-8393295c95fc-scripts\") pod \"placement-98dddfffb-l698f\" (UID: \"e913dfcd-dfe3-4c2c-b72a-8393295c95fc\") " pod="openstack/placement-98dddfffb-l698f" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.142609 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e913dfcd-dfe3-4c2c-b72a-8393295c95fc-combined-ca-bundle\") pod \"placement-98dddfffb-l698f\" (UID: \"e913dfcd-dfe3-4c2c-b72a-8393295c95fc\") " pod="openstack/placement-98dddfffb-l698f" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.142735 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e913dfcd-dfe3-4c2c-b72a-8393295c95fc-internal-tls-certs\") pod \"placement-98dddfffb-l698f\" (UID: \"e913dfcd-dfe3-4c2c-b72a-8393295c95fc\") " pod="openstack/placement-98dddfffb-l698f" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.244737 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v96lx\" (UniqueName: \"kubernetes.io/projected/e913dfcd-dfe3-4c2c-b72a-8393295c95fc-kube-api-access-v96lx\") pod \"placement-98dddfffb-l698f\" (UID: \"e913dfcd-dfe3-4c2c-b72a-8393295c95fc\") " pod="openstack/placement-98dddfffb-l698f" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.244932 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e913dfcd-dfe3-4c2c-b72a-8393295c95fc-logs\") pod \"placement-98dddfffb-l698f\" (UID: \"e913dfcd-dfe3-4c2c-b72a-8393295c95fc\") " pod="openstack/placement-98dddfffb-l698f" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.244984 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e913dfcd-dfe3-4c2c-b72a-8393295c95fc-config-data\") pod \"placement-98dddfffb-l698f\" (UID: \"e913dfcd-dfe3-4c2c-b72a-8393295c95fc\") " pod="openstack/placement-98dddfffb-l698f" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.245042 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e913dfcd-dfe3-4c2c-b72a-8393295c95fc-public-tls-certs\") pod \"placement-98dddfffb-l698f\" (UID: \"e913dfcd-dfe3-4c2c-b72a-8393295c95fc\") " pod="openstack/placement-98dddfffb-l698f" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.245153 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e913dfcd-dfe3-4c2c-b72a-8393295c95fc-scripts\") pod \"placement-98dddfffb-l698f\" (UID: \"e913dfcd-dfe3-4c2c-b72a-8393295c95fc\") " pod="openstack/placement-98dddfffb-l698f" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.245229 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e913dfcd-dfe3-4c2c-b72a-8393295c95fc-combined-ca-bundle\") pod \"placement-98dddfffb-l698f\" (UID: \"e913dfcd-dfe3-4c2c-b72a-8393295c95fc\") " pod="openstack/placement-98dddfffb-l698f" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.245286 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e913dfcd-dfe3-4c2c-b72a-8393295c95fc-internal-tls-certs\") pod \"placement-98dddfffb-l698f\" (UID: \"e913dfcd-dfe3-4c2c-b72a-8393295c95fc\") " pod="openstack/placement-98dddfffb-l698f" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.245342 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e913dfcd-dfe3-4c2c-b72a-8393295c95fc-logs\") pod \"placement-98dddfffb-l698f\" (UID: \"e913dfcd-dfe3-4c2c-b72a-8393295c95fc\") " pod="openstack/placement-98dddfffb-l698f" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.252732 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e913dfcd-dfe3-4c2c-b72a-8393295c95fc-public-tls-certs\") pod \"placement-98dddfffb-l698f\" (UID: \"e913dfcd-dfe3-4c2c-b72a-8393295c95fc\") " pod="openstack/placement-98dddfffb-l698f" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.252989 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e913dfcd-dfe3-4c2c-b72a-8393295c95fc-scripts\") pod \"placement-98dddfffb-l698f\" (UID: \"e913dfcd-dfe3-4c2c-b72a-8393295c95fc\") " pod="openstack/placement-98dddfffb-l698f" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.255251 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e913dfcd-dfe3-4c2c-b72a-8393295c95fc-combined-ca-bundle\") pod \"placement-98dddfffb-l698f\" (UID: \"e913dfcd-dfe3-4c2c-b72a-8393295c95fc\") " pod="openstack/placement-98dddfffb-l698f" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.255788 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e913dfcd-dfe3-4c2c-b72a-8393295c95fc-internal-tls-certs\") pod \"placement-98dddfffb-l698f\" (UID: \"e913dfcd-dfe3-4c2c-b72a-8393295c95fc\") " pod="openstack/placement-98dddfffb-l698f" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.259502 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e913dfcd-dfe3-4c2c-b72a-8393295c95fc-config-data\") pod \"placement-98dddfffb-l698f\" (UID: \"e913dfcd-dfe3-4c2c-b72a-8393295c95fc\") " pod="openstack/placement-98dddfffb-l698f" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.259878 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v96lx\" (UniqueName: \"kubernetes.io/projected/e913dfcd-dfe3-4c2c-b72a-8393295c95fc-kube-api-access-v96lx\") pod \"placement-98dddfffb-l698f\" (UID: \"e913dfcd-dfe3-4c2c-b72a-8393295c95fc\") " pod="openstack/placement-98dddfffb-l698f" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.358672 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-qk8sl" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.393334 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-98dddfffb-l698f" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.446879 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5c1f8568-2368-445f-a2cb-01ab3a6d1564-db-sync-config-data\") pod \"5c1f8568-2368-445f-a2cb-01ab3a6d1564\" (UID: \"5c1f8568-2368-445f-a2cb-01ab3a6d1564\") " Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.446982 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c1f8568-2368-445f-a2cb-01ab3a6d1564-combined-ca-bundle\") pod \"5c1f8568-2368-445f-a2cb-01ab3a6d1564\" (UID: \"5c1f8568-2368-445f-a2cb-01ab3a6d1564\") " Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.447109 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkchj\" (UniqueName: \"kubernetes.io/projected/5c1f8568-2368-445f-a2cb-01ab3a6d1564-kube-api-access-jkchj\") pod \"5c1f8568-2368-445f-a2cb-01ab3a6d1564\" (UID: \"5c1f8568-2368-445f-a2cb-01ab3a6d1564\") " Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.455946 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c1f8568-2368-445f-a2cb-01ab3a6d1564-kube-api-access-jkchj" (OuterVolumeSpecName: "kube-api-access-jkchj") pod "5c1f8568-2368-445f-a2cb-01ab3a6d1564" (UID: "5c1f8568-2368-445f-a2cb-01ab3a6d1564"). InnerVolumeSpecName "kube-api-access-jkchj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.459490 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c1f8568-2368-445f-a2cb-01ab3a6d1564-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "5c1f8568-2368-445f-a2cb-01ab3a6d1564" (UID: "5c1f8568-2368-445f-a2cb-01ab3a6d1564"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.495320 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c1f8568-2368-445f-a2cb-01ab3a6d1564-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5c1f8568-2368-445f-a2cb-01ab3a6d1564" (UID: "5c1f8568-2368-445f-a2cb-01ab3a6d1564"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.549618 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkchj\" (UniqueName: \"kubernetes.io/projected/5c1f8568-2368-445f-a2cb-01ab3a6d1564-kube-api-access-jkchj\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.549640 4813 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5c1f8568-2368-445f-a2cb-01ab3a6d1564-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.549649 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c1f8568-2368-445f-a2cb-01ab3a6d1564-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.895909 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-qk8sl" event={"ID":"5c1f8568-2368-445f-a2cb-01ab3a6d1564","Type":"ContainerDied","Data":"8d0dbd6457572bd8d3cecec5e9bc1befe436860af2eb9468111eb2dcaab364f9"} Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.896302 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d0dbd6457572bd8d3cecec5e9bc1befe436860af2eb9468111eb2dcaab364f9" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.896372 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-qk8sl" Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.899493 4813 generic.go:334] "Generic (PLEG): container finished" podID="67cf9ebe-1a58-4e50-9c10-f6f7933e781f" containerID="0a3cd73bf8fefc94389973058f01beb8ba5cf10cdc7fb883c3651d2a34f9ed30" exitCode=0 Dec 06 16:03:24 crc kubenswrapper[4813]: I1206 16:03:24.899599 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6qjfq" event={"ID":"67cf9ebe-1a58-4e50-9c10-f6f7933e781f","Type":"ContainerDied","Data":"0a3cd73bf8fefc94389973058f01beb8ba5cf10cdc7fb883c3651d2a34f9ed30"} Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.166356 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-ffc86b7c-26pxs"] Dec 06 16:03:25 crc kubenswrapper[4813]: E1206 16:03:25.166739 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c1f8568-2368-445f-a2cb-01ab3a6d1564" containerName="barbican-db-sync" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.166756 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c1f8568-2368-445f-a2cb-01ab3a6d1564" containerName="barbican-db-sync" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.166959 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c1f8568-2368-445f-a2cb-01ab3a6d1564" containerName="barbican-db-sync" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.167857 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-ffc86b7c-26pxs" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.172687 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-64449f75dd-k89r7"] Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.174071 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-64449f75dd-k89r7" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.176849 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.177092 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.177219 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-swv95" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.177366 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.201308 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-ffc86b7c-26pxs"] Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.233337 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-64449f75dd-k89r7"] Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.266677 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf88c989-e78b-456e-ba32-19d8af98a1d0-combined-ca-bundle\") pod \"barbican-worker-ffc86b7c-26pxs\" (UID: \"cf88c989-e78b-456e-ba32-19d8af98a1d0\") " pod="openstack/barbican-worker-ffc86b7c-26pxs" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.266742 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/322df0dd-86e6-466b-8afa-e4e52e889827-logs\") pod \"barbican-keystone-listener-64449f75dd-k89r7\" (UID: \"322df0dd-86e6-466b-8afa-e4e52e889827\") " pod="openstack/barbican-keystone-listener-64449f75dd-k89r7" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.266801 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/322df0dd-86e6-466b-8afa-e4e52e889827-config-data-custom\") pod \"barbican-keystone-listener-64449f75dd-k89r7\" (UID: \"322df0dd-86e6-466b-8afa-e4e52e889827\") " pod="openstack/barbican-keystone-listener-64449f75dd-k89r7" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.266870 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/322df0dd-86e6-466b-8afa-e4e52e889827-combined-ca-bundle\") pod \"barbican-keystone-listener-64449f75dd-k89r7\" (UID: \"322df0dd-86e6-466b-8afa-e4e52e889827\") " pod="openstack/barbican-keystone-listener-64449f75dd-k89r7" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.266898 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf88c989-e78b-456e-ba32-19d8af98a1d0-config-data\") pod \"barbican-worker-ffc86b7c-26pxs\" (UID: \"cf88c989-e78b-456e-ba32-19d8af98a1d0\") " pod="openstack/barbican-worker-ffc86b7c-26pxs" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.266920 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqgfp\" (UniqueName: \"kubernetes.io/projected/322df0dd-86e6-466b-8afa-e4e52e889827-kube-api-access-hqgfp\") pod \"barbican-keystone-listener-64449f75dd-k89r7\" (UID: \"322df0dd-86e6-466b-8afa-e4e52e889827\") " pod="openstack/barbican-keystone-listener-64449f75dd-k89r7" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.266963 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c24sd\" (UniqueName: \"kubernetes.io/projected/cf88c989-e78b-456e-ba32-19d8af98a1d0-kube-api-access-c24sd\") pod \"barbican-worker-ffc86b7c-26pxs\" (UID: \"cf88c989-e78b-456e-ba32-19d8af98a1d0\") " pod="openstack/barbican-worker-ffc86b7c-26pxs" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.266979 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cf88c989-e78b-456e-ba32-19d8af98a1d0-config-data-custom\") pod \"barbican-worker-ffc86b7c-26pxs\" (UID: \"cf88c989-e78b-456e-ba32-19d8af98a1d0\") " pod="openstack/barbican-worker-ffc86b7c-26pxs" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.266995 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/322df0dd-86e6-466b-8afa-e4e52e889827-config-data\") pod \"barbican-keystone-listener-64449f75dd-k89r7\" (UID: \"322df0dd-86e6-466b-8afa-e4e52e889827\") " pod="openstack/barbican-keystone-listener-64449f75dd-k89r7" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.267041 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf88c989-e78b-456e-ba32-19d8af98a1d0-logs\") pod \"barbican-worker-ffc86b7c-26pxs\" (UID: \"cf88c989-e78b-456e-ba32-19d8af98a1d0\") " pod="openstack/barbican-worker-ffc86b7c-26pxs" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.324804 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-qdtmw"] Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.325066 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" podUID="90124024-daf7-4241-a787-2840359fe074" containerName="dnsmasq-dns" containerID="cri-o://b4833ba9d04cff0f35659eeb3b203e748fe5c3214b9fed117882cb9e9972b2b8" gracePeriod=10 Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.326732 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.371645 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cf88c989-e78b-456e-ba32-19d8af98a1d0-config-data-custom\") pod \"barbican-worker-ffc86b7c-26pxs\" (UID: \"cf88c989-e78b-456e-ba32-19d8af98a1d0\") " pod="openstack/barbican-worker-ffc86b7c-26pxs" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.371683 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c24sd\" (UniqueName: \"kubernetes.io/projected/cf88c989-e78b-456e-ba32-19d8af98a1d0-kube-api-access-c24sd\") pod \"barbican-worker-ffc86b7c-26pxs\" (UID: \"cf88c989-e78b-456e-ba32-19d8af98a1d0\") " pod="openstack/barbican-worker-ffc86b7c-26pxs" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.371704 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/322df0dd-86e6-466b-8afa-e4e52e889827-config-data\") pod \"barbican-keystone-listener-64449f75dd-k89r7\" (UID: \"322df0dd-86e6-466b-8afa-e4e52e889827\") " pod="openstack/barbican-keystone-listener-64449f75dd-k89r7" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.371750 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf88c989-e78b-456e-ba32-19d8af98a1d0-logs\") pod \"barbican-worker-ffc86b7c-26pxs\" (UID: \"cf88c989-e78b-456e-ba32-19d8af98a1d0\") " pod="openstack/barbican-worker-ffc86b7c-26pxs" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.371774 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf88c989-e78b-456e-ba32-19d8af98a1d0-combined-ca-bundle\") pod \"barbican-worker-ffc86b7c-26pxs\" (UID: \"cf88c989-e78b-456e-ba32-19d8af98a1d0\") " pod="openstack/barbican-worker-ffc86b7c-26pxs" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.371791 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/322df0dd-86e6-466b-8afa-e4e52e889827-logs\") pod \"barbican-keystone-listener-64449f75dd-k89r7\" (UID: \"322df0dd-86e6-466b-8afa-e4e52e889827\") " pod="openstack/barbican-keystone-listener-64449f75dd-k89r7" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.371818 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/322df0dd-86e6-466b-8afa-e4e52e889827-config-data-custom\") pod \"barbican-keystone-listener-64449f75dd-k89r7\" (UID: \"322df0dd-86e6-466b-8afa-e4e52e889827\") " pod="openstack/barbican-keystone-listener-64449f75dd-k89r7" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.371864 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/322df0dd-86e6-466b-8afa-e4e52e889827-combined-ca-bundle\") pod \"barbican-keystone-listener-64449f75dd-k89r7\" (UID: \"322df0dd-86e6-466b-8afa-e4e52e889827\") " pod="openstack/barbican-keystone-listener-64449f75dd-k89r7" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.371888 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf88c989-e78b-456e-ba32-19d8af98a1d0-config-data\") pod \"barbican-worker-ffc86b7c-26pxs\" (UID: \"cf88c989-e78b-456e-ba32-19d8af98a1d0\") " pod="openstack/barbican-worker-ffc86b7c-26pxs" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.371906 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqgfp\" (UniqueName: \"kubernetes.io/projected/322df0dd-86e6-466b-8afa-e4e52e889827-kube-api-access-hqgfp\") pod \"barbican-keystone-listener-64449f75dd-k89r7\" (UID: \"322df0dd-86e6-466b-8afa-e4e52e889827\") " pod="openstack/barbican-keystone-listener-64449f75dd-k89r7" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.373369 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/322df0dd-86e6-466b-8afa-e4e52e889827-logs\") pod \"barbican-keystone-listener-64449f75dd-k89r7\" (UID: \"322df0dd-86e6-466b-8afa-e4e52e889827\") " pod="openstack/barbican-keystone-listener-64449f75dd-k89r7" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.374138 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-lnl2v"] Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.377649 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/322df0dd-86e6-466b-8afa-e4e52e889827-config-data\") pod \"barbican-keystone-listener-64449f75dd-k89r7\" (UID: \"322df0dd-86e6-466b-8afa-e4e52e889827\") " pod="openstack/barbican-keystone-listener-64449f75dd-k89r7" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.377826 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf88c989-e78b-456e-ba32-19d8af98a1d0-logs\") pod \"barbican-worker-ffc86b7c-26pxs\" (UID: \"cf88c989-e78b-456e-ba32-19d8af98a1d0\") " pod="openstack/barbican-worker-ffc86b7c-26pxs" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.414595 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.416889 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf88c989-e78b-456e-ba32-19d8af98a1d0-combined-ca-bundle\") pod \"barbican-worker-ffc86b7c-26pxs\" (UID: \"cf88c989-e78b-456e-ba32-19d8af98a1d0\") " pod="openstack/barbican-worker-ffc86b7c-26pxs" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.429800 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf88c989-e78b-456e-ba32-19d8af98a1d0-config-data\") pod \"barbican-worker-ffc86b7c-26pxs\" (UID: \"cf88c989-e78b-456e-ba32-19d8af98a1d0\") " pod="openstack/barbican-worker-ffc86b7c-26pxs" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.431515 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/322df0dd-86e6-466b-8afa-e4e52e889827-combined-ca-bundle\") pod \"barbican-keystone-listener-64449f75dd-k89r7\" (UID: \"322df0dd-86e6-466b-8afa-e4e52e889827\") " pod="openstack/barbican-keystone-listener-64449f75dd-k89r7" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.431968 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cf88c989-e78b-456e-ba32-19d8af98a1d0-config-data-custom\") pod \"barbican-worker-ffc86b7c-26pxs\" (UID: \"cf88c989-e78b-456e-ba32-19d8af98a1d0\") " pod="openstack/barbican-worker-ffc86b7c-26pxs" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.452246 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/322df0dd-86e6-466b-8afa-e4e52e889827-config-data-custom\") pod \"barbican-keystone-listener-64449f75dd-k89r7\" (UID: \"322df0dd-86e6-466b-8afa-e4e52e889827\") " pod="openstack/barbican-keystone-listener-64449f75dd-k89r7" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.476805 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-lnl2v"] Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.477376 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/264bf858-a126-4005-9c98-0515834c4956-config\") pod \"dnsmasq-dns-848cf88cfc-lnl2v\" (UID: \"264bf858-a126-4005-9c98-0515834c4956\") " pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.477415 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/264bf858-a126-4005-9c98-0515834c4956-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-lnl2v\" (UID: \"264bf858-a126-4005-9c98-0515834c4956\") " pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.477525 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6796z\" (UniqueName: \"kubernetes.io/projected/264bf858-a126-4005-9c98-0515834c4956-kube-api-access-6796z\") pod \"dnsmasq-dns-848cf88cfc-lnl2v\" (UID: \"264bf858-a126-4005-9c98-0515834c4956\") " pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.477578 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/264bf858-a126-4005-9c98-0515834c4956-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-lnl2v\" (UID: \"264bf858-a126-4005-9c98-0515834c4956\") " pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.477608 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/264bf858-a126-4005-9c98-0515834c4956-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-lnl2v\" (UID: \"264bf858-a126-4005-9c98-0515834c4956\") " pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.477638 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/264bf858-a126-4005-9c98-0515834c4956-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-lnl2v\" (UID: \"264bf858-a126-4005-9c98-0515834c4956\") " pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.479556 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c24sd\" (UniqueName: \"kubernetes.io/projected/cf88c989-e78b-456e-ba32-19d8af98a1d0-kube-api-access-c24sd\") pod \"barbican-worker-ffc86b7c-26pxs\" (UID: \"cf88c989-e78b-456e-ba32-19d8af98a1d0\") " pod="openstack/barbican-worker-ffc86b7c-26pxs" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.484556 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqgfp\" (UniqueName: \"kubernetes.io/projected/322df0dd-86e6-466b-8afa-e4e52e889827-kube-api-access-hqgfp\") pod \"barbican-keystone-listener-64449f75dd-k89r7\" (UID: \"322df0dd-86e6-466b-8afa-e4e52e889827\") " pod="openstack/barbican-keystone-listener-64449f75dd-k89r7" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.519066 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-64449f75dd-k89r7" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.519646 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-ffc86b7c-26pxs" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.582090 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6796z\" (UniqueName: \"kubernetes.io/projected/264bf858-a126-4005-9c98-0515834c4956-kube-api-access-6796z\") pod \"dnsmasq-dns-848cf88cfc-lnl2v\" (UID: \"264bf858-a126-4005-9c98-0515834c4956\") " pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.582150 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/264bf858-a126-4005-9c98-0515834c4956-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-lnl2v\" (UID: \"264bf858-a126-4005-9c98-0515834c4956\") " pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.582172 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/264bf858-a126-4005-9c98-0515834c4956-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-lnl2v\" (UID: \"264bf858-a126-4005-9c98-0515834c4956\") " pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.582200 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/264bf858-a126-4005-9c98-0515834c4956-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-lnl2v\" (UID: \"264bf858-a126-4005-9c98-0515834c4956\") " pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.582236 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/264bf858-a126-4005-9c98-0515834c4956-config\") pod \"dnsmasq-dns-848cf88cfc-lnl2v\" (UID: \"264bf858-a126-4005-9c98-0515834c4956\") " pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.582254 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/264bf858-a126-4005-9c98-0515834c4956-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-lnl2v\" (UID: \"264bf858-a126-4005-9c98-0515834c4956\") " pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.583232 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/264bf858-a126-4005-9c98-0515834c4956-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-lnl2v\" (UID: \"264bf858-a126-4005-9c98-0515834c4956\") " pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.583936 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/264bf858-a126-4005-9c98-0515834c4956-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-lnl2v\" (UID: \"264bf858-a126-4005-9c98-0515834c4956\") " pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.586934 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/264bf858-a126-4005-9c98-0515834c4956-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-lnl2v\" (UID: \"264bf858-a126-4005-9c98-0515834c4956\") " pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.587742 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/264bf858-a126-4005-9c98-0515834c4956-config\") pod \"dnsmasq-dns-848cf88cfc-lnl2v\" (UID: \"264bf858-a126-4005-9c98-0515834c4956\") " pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.592545 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/264bf858-a126-4005-9c98-0515834c4956-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-lnl2v\" (UID: \"264bf858-a126-4005-9c98-0515834c4956\") " pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.634783 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6796z\" (UniqueName: \"kubernetes.io/projected/264bf858-a126-4005-9c98-0515834c4956-kube-api-access-6796z\") pod \"dnsmasq-dns-848cf88cfc-lnl2v\" (UID: \"264bf858-a126-4005-9c98-0515834c4956\") " pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.640585 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.674788 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-66f946f97d-gh5xc"] Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.676079 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-66f946f97d-gh5xc" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.679450 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.700943 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-66f946f97d-gh5xc"] Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.790451 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78cfb04c-4099-4211-a82e-47a304251ab5-config-data\") pod \"barbican-api-66f946f97d-gh5xc\" (UID: \"78cfb04c-4099-4211-a82e-47a304251ab5\") " pod="openstack/barbican-api-66f946f97d-gh5xc" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.790501 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78cfb04c-4099-4211-a82e-47a304251ab5-combined-ca-bundle\") pod \"barbican-api-66f946f97d-gh5xc\" (UID: \"78cfb04c-4099-4211-a82e-47a304251ab5\") " pod="openstack/barbican-api-66f946f97d-gh5xc" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.790549 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78cfb04c-4099-4211-a82e-47a304251ab5-config-data-custom\") pod \"barbican-api-66f946f97d-gh5xc\" (UID: \"78cfb04c-4099-4211-a82e-47a304251ab5\") " pod="openstack/barbican-api-66f946f97d-gh5xc" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.790577 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78cfb04c-4099-4211-a82e-47a304251ab5-logs\") pod \"barbican-api-66f946f97d-gh5xc\" (UID: \"78cfb04c-4099-4211-a82e-47a304251ab5\") " pod="openstack/barbican-api-66f946f97d-gh5xc" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.790609 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bwn7\" (UniqueName: \"kubernetes.io/projected/78cfb04c-4099-4211-a82e-47a304251ab5-kube-api-access-6bwn7\") pod \"barbican-api-66f946f97d-gh5xc\" (UID: \"78cfb04c-4099-4211-a82e-47a304251ab5\") " pod="openstack/barbican-api-66f946f97d-gh5xc" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.892339 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78cfb04c-4099-4211-a82e-47a304251ab5-config-data\") pod \"barbican-api-66f946f97d-gh5xc\" (UID: \"78cfb04c-4099-4211-a82e-47a304251ab5\") " pod="openstack/barbican-api-66f946f97d-gh5xc" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.892389 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78cfb04c-4099-4211-a82e-47a304251ab5-combined-ca-bundle\") pod \"barbican-api-66f946f97d-gh5xc\" (UID: \"78cfb04c-4099-4211-a82e-47a304251ab5\") " pod="openstack/barbican-api-66f946f97d-gh5xc" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.892452 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78cfb04c-4099-4211-a82e-47a304251ab5-config-data-custom\") pod \"barbican-api-66f946f97d-gh5xc\" (UID: \"78cfb04c-4099-4211-a82e-47a304251ab5\") " pod="openstack/barbican-api-66f946f97d-gh5xc" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.892482 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78cfb04c-4099-4211-a82e-47a304251ab5-logs\") pod \"barbican-api-66f946f97d-gh5xc\" (UID: \"78cfb04c-4099-4211-a82e-47a304251ab5\") " pod="openstack/barbican-api-66f946f97d-gh5xc" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.892515 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bwn7\" (UniqueName: \"kubernetes.io/projected/78cfb04c-4099-4211-a82e-47a304251ab5-kube-api-access-6bwn7\") pod \"barbican-api-66f946f97d-gh5xc\" (UID: \"78cfb04c-4099-4211-a82e-47a304251ab5\") " pod="openstack/barbican-api-66f946f97d-gh5xc" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.893296 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78cfb04c-4099-4211-a82e-47a304251ab5-logs\") pod \"barbican-api-66f946f97d-gh5xc\" (UID: \"78cfb04c-4099-4211-a82e-47a304251ab5\") " pod="openstack/barbican-api-66f946f97d-gh5xc" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.898846 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78cfb04c-4099-4211-a82e-47a304251ab5-combined-ca-bundle\") pod \"barbican-api-66f946f97d-gh5xc\" (UID: \"78cfb04c-4099-4211-a82e-47a304251ab5\") " pod="openstack/barbican-api-66f946f97d-gh5xc" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.902022 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78cfb04c-4099-4211-a82e-47a304251ab5-config-data-custom\") pod \"barbican-api-66f946f97d-gh5xc\" (UID: \"78cfb04c-4099-4211-a82e-47a304251ab5\") " pod="openstack/barbican-api-66f946f97d-gh5xc" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.911906 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78cfb04c-4099-4211-a82e-47a304251ab5-config-data\") pod \"barbican-api-66f946f97d-gh5xc\" (UID: \"78cfb04c-4099-4211-a82e-47a304251ab5\") " pod="openstack/barbican-api-66f946f97d-gh5xc" Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.926196 4813 generic.go:334] "Generic (PLEG): container finished" podID="90124024-daf7-4241-a787-2840359fe074" containerID="b4833ba9d04cff0f35659eeb3b203e748fe5c3214b9fed117882cb9e9972b2b8" exitCode=0 Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.926393 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" event={"ID":"90124024-daf7-4241-a787-2840359fe074","Type":"ContainerDied","Data":"b4833ba9d04cff0f35659eeb3b203e748fe5c3214b9fed117882cb9e9972b2b8"} Dec 06 16:03:25 crc kubenswrapper[4813]: I1206 16:03:25.927967 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bwn7\" (UniqueName: \"kubernetes.io/projected/78cfb04c-4099-4211-a82e-47a304251ab5-kube-api-access-6bwn7\") pod \"barbican-api-66f946f97d-gh5xc\" (UID: \"78cfb04c-4099-4211-a82e-47a304251ab5\") " pod="openstack/barbican-api-66f946f97d-gh5xc" Dec 06 16:03:26 crc kubenswrapper[4813]: I1206 16:03:26.014185 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-66f946f97d-gh5xc" Dec 06 16:03:26 crc kubenswrapper[4813]: I1206 16:03:26.231553 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-bbc668d58-h8w67" Dec 06 16:03:26 crc kubenswrapper[4813]: I1206 16:03:26.231618 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-bbc668d58-h8w67" Dec 06 16:03:26 crc kubenswrapper[4813]: I1206 16:03:26.344005 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7fd6f7946b-kmm96" Dec 06 16:03:26 crc kubenswrapper[4813]: I1206 16:03:26.344057 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7fd6f7946b-kmm96" Dec 06 16:03:29 crc kubenswrapper[4813]: I1206 16:03:29.126493 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-58d4d548d-6bl8w"] Dec 06 16:03:29 crc kubenswrapper[4813]: I1206 16:03:29.128155 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-58d4d548d-6bl8w" Dec 06 16:03:29 crc kubenswrapper[4813]: I1206 16:03:29.131607 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 06 16:03:29 crc kubenswrapper[4813]: I1206 16:03:29.131753 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 06 16:03:29 crc kubenswrapper[4813]: I1206 16:03:29.141022 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-58d4d548d-6bl8w"] Dec 06 16:03:29 crc kubenswrapper[4813]: I1206 16:03:29.268103 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f883aaa3-f010-4c52-bc12-1591fb0ddb4d-internal-tls-certs\") pod \"barbican-api-58d4d548d-6bl8w\" (UID: \"f883aaa3-f010-4c52-bc12-1591fb0ddb4d\") " pod="openstack/barbican-api-58d4d548d-6bl8w" Dec 06 16:03:29 crc kubenswrapper[4813]: I1206 16:03:29.268181 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f883aaa3-f010-4c52-bc12-1591fb0ddb4d-config-data-custom\") pod \"barbican-api-58d4d548d-6bl8w\" (UID: \"f883aaa3-f010-4c52-bc12-1591fb0ddb4d\") " pod="openstack/barbican-api-58d4d548d-6bl8w" Dec 06 16:03:29 crc kubenswrapper[4813]: I1206 16:03:29.268227 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f883aaa3-f010-4c52-bc12-1591fb0ddb4d-config-data\") pod \"barbican-api-58d4d548d-6bl8w\" (UID: \"f883aaa3-f010-4c52-bc12-1591fb0ddb4d\") " pod="openstack/barbican-api-58d4d548d-6bl8w" Dec 06 16:03:29 crc kubenswrapper[4813]: I1206 16:03:29.268255 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f883aaa3-f010-4c52-bc12-1591fb0ddb4d-public-tls-certs\") pod \"barbican-api-58d4d548d-6bl8w\" (UID: \"f883aaa3-f010-4c52-bc12-1591fb0ddb4d\") " pod="openstack/barbican-api-58d4d548d-6bl8w" Dec 06 16:03:29 crc kubenswrapper[4813]: I1206 16:03:29.268449 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gd92c\" (UniqueName: \"kubernetes.io/projected/f883aaa3-f010-4c52-bc12-1591fb0ddb4d-kube-api-access-gd92c\") pod \"barbican-api-58d4d548d-6bl8w\" (UID: \"f883aaa3-f010-4c52-bc12-1591fb0ddb4d\") " pod="openstack/barbican-api-58d4d548d-6bl8w" Dec 06 16:03:29 crc kubenswrapper[4813]: I1206 16:03:29.268509 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f883aaa3-f010-4c52-bc12-1591fb0ddb4d-combined-ca-bundle\") pod \"barbican-api-58d4d548d-6bl8w\" (UID: \"f883aaa3-f010-4c52-bc12-1591fb0ddb4d\") " pod="openstack/barbican-api-58d4d548d-6bl8w" Dec 06 16:03:29 crc kubenswrapper[4813]: I1206 16:03:29.268626 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f883aaa3-f010-4c52-bc12-1591fb0ddb4d-logs\") pod \"barbican-api-58d4d548d-6bl8w\" (UID: \"f883aaa3-f010-4c52-bc12-1591fb0ddb4d\") " pod="openstack/barbican-api-58d4d548d-6bl8w" Dec 06 16:03:29 crc kubenswrapper[4813]: I1206 16:03:29.369777 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f883aaa3-f010-4c52-bc12-1591fb0ddb4d-config-data-custom\") pod \"barbican-api-58d4d548d-6bl8w\" (UID: \"f883aaa3-f010-4c52-bc12-1591fb0ddb4d\") " pod="openstack/barbican-api-58d4d548d-6bl8w" Dec 06 16:03:29 crc kubenswrapper[4813]: I1206 16:03:29.369836 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f883aaa3-f010-4c52-bc12-1591fb0ddb4d-config-data\") pod \"barbican-api-58d4d548d-6bl8w\" (UID: \"f883aaa3-f010-4c52-bc12-1591fb0ddb4d\") " pod="openstack/barbican-api-58d4d548d-6bl8w" Dec 06 16:03:29 crc kubenswrapper[4813]: I1206 16:03:29.369865 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f883aaa3-f010-4c52-bc12-1591fb0ddb4d-public-tls-certs\") pod \"barbican-api-58d4d548d-6bl8w\" (UID: \"f883aaa3-f010-4c52-bc12-1591fb0ddb4d\") " pod="openstack/barbican-api-58d4d548d-6bl8w" Dec 06 16:03:29 crc kubenswrapper[4813]: I1206 16:03:29.369918 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gd92c\" (UniqueName: \"kubernetes.io/projected/f883aaa3-f010-4c52-bc12-1591fb0ddb4d-kube-api-access-gd92c\") pod \"barbican-api-58d4d548d-6bl8w\" (UID: \"f883aaa3-f010-4c52-bc12-1591fb0ddb4d\") " pod="openstack/barbican-api-58d4d548d-6bl8w" Dec 06 16:03:29 crc kubenswrapper[4813]: I1206 16:03:29.369937 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f883aaa3-f010-4c52-bc12-1591fb0ddb4d-combined-ca-bundle\") pod \"barbican-api-58d4d548d-6bl8w\" (UID: \"f883aaa3-f010-4c52-bc12-1591fb0ddb4d\") " pod="openstack/barbican-api-58d4d548d-6bl8w" Dec 06 16:03:29 crc kubenswrapper[4813]: I1206 16:03:29.369979 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f883aaa3-f010-4c52-bc12-1591fb0ddb4d-logs\") pod \"barbican-api-58d4d548d-6bl8w\" (UID: \"f883aaa3-f010-4c52-bc12-1591fb0ddb4d\") " pod="openstack/barbican-api-58d4d548d-6bl8w" Dec 06 16:03:29 crc kubenswrapper[4813]: I1206 16:03:29.370027 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f883aaa3-f010-4c52-bc12-1591fb0ddb4d-internal-tls-certs\") pod \"barbican-api-58d4d548d-6bl8w\" (UID: \"f883aaa3-f010-4c52-bc12-1591fb0ddb4d\") " pod="openstack/barbican-api-58d4d548d-6bl8w" Dec 06 16:03:29 crc kubenswrapper[4813]: I1206 16:03:29.372749 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f883aaa3-f010-4c52-bc12-1591fb0ddb4d-logs\") pod \"barbican-api-58d4d548d-6bl8w\" (UID: \"f883aaa3-f010-4c52-bc12-1591fb0ddb4d\") " pod="openstack/barbican-api-58d4d548d-6bl8w" Dec 06 16:03:29 crc kubenswrapper[4813]: I1206 16:03:29.377013 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f883aaa3-f010-4c52-bc12-1591fb0ddb4d-internal-tls-certs\") pod \"barbican-api-58d4d548d-6bl8w\" (UID: \"f883aaa3-f010-4c52-bc12-1591fb0ddb4d\") " pod="openstack/barbican-api-58d4d548d-6bl8w" Dec 06 16:03:29 crc kubenswrapper[4813]: I1206 16:03:29.377652 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f883aaa3-f010-4c52-bc12-1591fb0ddb4d-config-data\") pod \"barbican-api-58d4d548d-6bl8w\" (UID: \"f883aaa3-f010-4c52-bc12-1591fb0ddb4d\") " pod="openstack/barbican-api-58d4d548d-6bl8w" Dec 06 16:03:29 crc kubenswrapper[4813]: I1206 16:03:29.381832 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f883aaa3-f010-4c52-bc12-1591fb0ddb4d-combined-ca-bundle\") pod \"barbican-api-58d4d548d-6bl8w\" (UID: \"f883aaa3-f010-4c52-bc12-1591fb0ddb4d\") " pod="openstack/barbican-api-58d4d548d-6bl8w" Dec 06 16:03:29 crc kubenswrapper[4813]: I1206 16:03:29.384996 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f883aaa3-f010-4c52-bc12-1591fb0ddb4d-public-tls-certs\") pod \"barbican-api-58d4d548d-6bl8w\" (UID: \"f883aaa3-f010-4c52-bc12-1591fb0ddb4d\") " pod="openstack/barbican-api-58d4d548d-6bl8w" Dec 06 16:03:29 crc kubenswrapper[4813]: I1206 16:03:29.385494 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f883aaa3-f010-4c52-bc12-1591fb0ddb4d-config-data-custom\") pod \"barbican-api-58d4d548d-6bl8w\" (UID: \"f883aaa3-f010-4c52-bc12-1591fb0ddb4d\") " pod="openstack/barbican-api-58d4d548d-6bl8w" Dec 06 16:03:29 crc kubenswrapper[4813]: I1206 16:03:29.395900 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gd92c\" (UniqueName: \"kubernetes.io/projected/f883aaa3-f010-4c52-bc12-1591fb0ddb4d-kube-api-access-gd92c\") pod \"barbican-api-58d4d548d-6bl8w\" (UID: \"f883aaa3-f010-4c52-bc12-1591fb0ddb4d\") " pod="openstack/barbican-api-58d4d548d-6bl8w" Dec 06 16:03:29 crc kubenswrapper[4813]: I1206 16:03:29.444786 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-58d4d548d-6bl8w" Dec 06 16:03:30 crc kubenswrapper[4813]: I1206 16:03:30.193771 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 06 16:03:30 crc kubenswrapper[4813]: I1206 16:03:30.194099 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 06 16:03:30 crc kubenswrapper[4813]: I1206 16:03:30.246879 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 06 16:03:30 crc kubenswrapper[4813]: I1206 16:03:30.277684 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 06 16:03:30 crc kubenswrapper[4813]: I1206 16:03:30.347094 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" podUID="90124024-daf7-4241-a787-2840359fe074" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.151:5353: connect: connection refused" Dec 06 16:03:30 crc kubenswrapper[4813]: I1206 16:03:30.824328 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 06 16:03:30 crc kubenswrapper[4813]: I1206 16:03:30.824537 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 06 16:03:30 crc kubenswrapper[4813]: I1206 16:03:30.865919 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 06 16:03:30 crc kubenswrapper[4813]: I1206 16:03:30.876870 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 06 16:03:30 crc kubenswrapper[4813]: I1206 16:03:30.965082 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 06 16:03:30 crc kubenswrapper[4813]: I1206 16:03:30.967541 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 06 16:03:30 crc kubenswrapper[4813]: I1206 16:03:30.967569 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 06 16:03:30 crc kubenswrapper[4813]: I1206 16:03:30.967579 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 06 16:03:31 crc kubenswrapper[4813]: I1206 16:03:31.521853 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6qjfq" Dec 06 16:03:31 crc kubenswrapper[4813]: I1206 16:03:31.610824 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-combined-ca-bundle\") pod \"67cf9ebe-1a58-4e50-9c10-f6f7933e781f\" (UID: \"67cf9ebe-1a58-4e50-9c10-f6f7933e781f\") " Dec 06 16:03:31 crc kubenswrapper[4813]: I1206 16:03:31.610894 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-455vp\" (UniqueName: \"kubernetes.io/projected/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-kube-api-access-455vp\") pod \"67cf9ebe-1a58-4e50-9c10-f6f7933e781f\" (UID: \"67cf9ebe-1a58-4e50-9c10-f6f7933e781f\") " Dec 06 16:03:31 crc kubenswrapper[4813]: I1206 16:03:31.610999 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-fernet-keys\") pod \"67cf9ebe-1a58-4e50-9c10-f6f7933e781f\" (UID: \"67cf9ebe-1a58-4e50-9c10-f6f7933e781f\") " Dec 06 16:03:31 crc kubenswrapper[4813]: I1206 16:03:31.611044 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-scripts\") pod \"67cf9ebe-1a58-4e50-9c10-f6f7933e781f\" (UID: \"67cf9ebe-1a58-4e50-9c10-f6f7933e781f\") " Dec 06 16:03:31 crc kubenswrapper[4813]: I1206 16:03:31.611064 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-config-data\") pod \"67cf9ebe-1a58-4e50-9c10-f6f7933e781f\" (UID: \"67cf9ebe-1a58-4e50-9c10-f6f7933e781f\") " Dec 06 16:03:31 crc kubenswrapper[4813]: I1206 16:03:31.611080 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-credential-keys\") pod \"67cf9ebe-1a58-4e50-9c10-f6f7933e781f\" (UID: \"67cf9ebe-1a58-4e50-9c10-f6f7933e781f\") " Dec 06 16:03:31 crc kubenswrapper[4813]: I1206 16:03:31.640286 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "67cf9ebe-1a58-4e50-9c10-f6f7933e781f" (UID: "67cf9ebe-1a58-4e50-9c10-f6f7933e781f"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:31 crc kubenswrapper[4813]: I1206 16:03:31.641126 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "67cf9ebe-1a58-4e50-9c10-f6f7933e781f" (UID: "67cf9ebe-1a58-4e50-9c10-f6f7933e781f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:31 crc kubenswrapper[4813]: I1206 16:03:31.641892 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-kube-api-access-455vp" (OuterVolumeSpecName: "kube-api-access-455vp") pod "67cf9ebe-1a58-4e50-9c10-f6f7933e781f" (UID: "67cf9ebe-1a58-4e50-9c10-f6f7933e781f"). InnerVolumeSpecName "kube-api-access-455vp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:03:31 crc kubenswrapper[4813]: I1206 16:03:31.646461 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-scripts" (OuterVolumeSpecName: "scripts") pod "67cf9ebe-1a58-4e50-9c10-f6f7933e781f" (UID: "67cf9ebe-1a58-4e50-9c10-f6f7933e781f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:31 crc kubenswrapper[4813]: I1206 16:03:31.676406 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-config-data" (OuterVolumeSpecName: "config-data") pod "67cf9ebe-1a58-4e50-9c10-f6f7933e781f" (UID: "67cf9ebe-1a58-4e50-9c10-f6f7933e781f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:31 crc kubenswrapper[4813]: I1206 16:03:31.708407 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67cf9ebe-1a58-4e50-9c10-f6f7933e781f" (UID: "67cf9ebe-1a58-4e50-9c10-f6f7933e781f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:31 crc kubenswrapper[4813]: I1206 16:03:31.717482 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:31 crc kubenswrapper[4813]: I1206 16:03:31.717512 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:31 crc kubenswrapper[4813]: I1206 16:03:31.717522 4813 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:31 crc kubenswrapper[4813]: I1206 16:03:31.717533 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:31 crc kubenswrapper[4813]: I1206 16:03:31.717541 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-455vp\" (UniqueName: \"kubernetes.io/projected/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-kube-api-access-455vp\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:31 crc kubenswrapper[4813]: I1206 16:03:31.717549 4813 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/67cf9ebe-1a58-4e50-9c10-f6f7933e781f-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.007986 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6qjfq" event={"ID":"67cf9ebe-1a58-4e50-9c10-f6f7933e781f","Type":"ContainerDied","Data":"4b25e79d792b968cd8c068aafbe3d9c04158ce7f04d3054381534631e83b5490"} Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.008226 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6qjfq" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.008271 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b25e79d792b968cd8c068aafbe3d9c04158ce7f04d3054381534631e83b5490" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.015472 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.124835 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2tb4\" (UniqueName: \"kubernetes.io/projected/90124024-daf7-4241-a787-2840359fe074-kube-api-access-c2tb4\") pod \"90124024-daf7-4241-a787-2840359fe074\" (UID: \"90124024-daf7-4241-a787-2840359fe074\") " Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.125123 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90124024-daf7-4241-a787-2840359fe074-dns-svc\") pod \"90124024-daf7-4241-a787-2840359fe074\" (UID: \"90124024-daf7-4241-a787-2840359fe074\") " Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.125201 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/90124024-daf7-4241-a787-2840359fe074-dns-swift-storage-0\") pod \"90124024-daf7-4241-a787-2840359fe074\" (UID: \"90124024-daf7-4241-a787-2840359fe074\") " Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.125434 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90124024-daf7-4241-a787-2840359fe074-config\") pod \"90124024-daf7-4241-a787-2840359fe074\" (UID: \"90124024-daf7-4241-a787-2840359fe074\") " Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.125519 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/90124024-daf7-4241-a787-2840359fe074-ovsdbserver-sb\") pod \"90124024-daf7-4241-a787-2840359fe074\" (UID: \"90124024-daf7-4241-a787-2840359fe074\") " Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.125592 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/90124024-daf7-4241-a787-2840359fe074-ovsdbserver-nb\") pod \"90124024-daf7-4241-a787-2840359fe074\" (UID: \"90124024-daf7-4241-a787-2840359fe074\") " Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.127711 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90124024-daf7-4241-a787-2840359fe074-kube-api-access-c2tb4" (OuterVolumeSpecName: "kube-api-access-c2tb4") pod "90124024-daf7-4241-a787-2840359fe074" (UID: "90124024-daf7-4241-a787-2840359fe074"). InnerVolumeSpecName "kube-api-access-c2tb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.233436 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2tb4\" (UniqueName: \"kubernetes.io/projected/90124024-daf7-4241-a787-2840359fe074-kube-api-access-c2tb4\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.319692 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90124024-daf7-4241-a787-2840359fe074-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "90124024-daf7-4241-a787-2840359fe074" (UID: "90124024-daf7-4241-a787-2840359fe074"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.319848 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90124024-daf7-4241-a787-2840359fe074-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "90124024-daf7-4241-a787-2840359fe074" (UID: "90124024-daf7-4241-a787-2840359fe074"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.336804 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/90124024-daf7-4241-a787-2840359fe074-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.337109 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90124024-daf7-4241-a787-2840359fe074-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.361613 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90124024-daf7-4241-a787-2840359fe074-config" (OuterVolumeSpecName: "config") pod "90124024-daf7-4241-a787-2840359fe074" (UID: "90124024-daf7-4241-a787-2840359fe074"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.364734 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90124024-daf7-4241-a787-2840359fe074-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "90124024-daf7-4241-a787-2840359fe074" (UID: "90124024-daf7-4241-a787-2840359fe074"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.375640 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90124024-daf7-4241-a787-2840359fe074-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "90124024-daf7-4241-a787-2840359fe074" (UID: "90124024-daf7-4241-a787-2840359fe074"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.438479 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90124024-daf7-4241-a787-2840359fe074-config\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.438510 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/90124024-daf7-4241-a787-2840359fe074-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.438519 4813 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/90124024-daf7-4241-a787-2840359fe074-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.538803 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-lnl2v"] Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.734023 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-66f946f97d-gh5xc"] Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.762295 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-64449f75dd-k89r7"] Dec 06 16:03:32 crc kubenswrapper[4813]: W1206 16:03:32.869518 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf883aaa3_f010_4c52_bc12_1591fb0ddb4d.slice/crio-297489db702f144b7e8d1f1e9200a74e4aa43293ea6acc847fe523ace3575f60 WatchSource:0}: Error finding container 297489db702f144b7e8d1f1e9200a74e4aa43293ea6acc847fe523ace3575f60: Status 404 returned error can't find the container with id 297489db702f144b7e8d1f1e9200a74e4aa43293ea6acc847fe523ace3575f60 Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.895079 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-c4f8d8d9b-p5pqq"] Dec 06 16:03:32 crc kubenswrapper[4813]: E1206 16:03:32.895515 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67cf9ebe-1a58-4e50-9c10-f6f7933e781f" containerName="keystone-bootstrap" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.895528 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="67cf9ebe-1a58-4e50-9c10-f6f7933e781f" containerName="keystone-bootstrap" Dec 06 16:03:32 crc kubenswrapper[4813]: E1206 16:03:32.895552 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90124024-daf7-4241-a787-2840359fe074" containerName="init" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.895558 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="90124024-daf7-4241-a787-2840359fe074" containerName="init" Dec 06 16:03:32 crc kubenswrapper[4813]: E1206 16:03:32.895573 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90124024-daf7-4241-a787-2840359fe074" containerName="dnsmasq-dns" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.895579 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="90124024-daf7-4241-a787-2840359fe074" containerName="dnsmasq-dns" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.895738 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="90124024-daf7-4241-a787-2840359fe074" containerName="dnsmasq-dns" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.895752 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="67cf9ebe-1a58-4e50-9c10-f6f7933e781f" containerName="keystone-bootstrap" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.897113 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c4f8d8d9b-p5pqq" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.901282 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fmvk2" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.901413 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.901480 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.902428 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.910358 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.910513 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.926132 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-98dddfffb-l698f"] Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.944508 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-c4f8d8d9b-p5pqq"] Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.944540 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-58d4d548d-6bl8w"] Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.984516 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6d636bb-c76a-4895-8af6-6801f45e4e5b-public-tls-certs\") pod \"keystone-c4f8d8d9b-p5pqq\" (UID: \"c6d636bb-c76a-4895-8af6-6801f45e4e5b\") " pod="openstack/keystone-c4f8d8d9b-p5pqq" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.984607 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c6d636bb-c76a-4895-8af6-6801f45e4e5b-credential-keys\") pod \"keystone-c4f8d8d9b-p5pqq\" (UID: \"c6d636bb-c76a-4895-8af6-6801f45e4e5b\") " pod="openstack/keystone-c4f8d8d9b-p5pqq" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.984637 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6d636bb-c76a-4895-8af6-6801f45e4e5b-internal-tls-certs\") pod \"keystone-c4f8d8d9b-p5pqq\" (UID: \"c6d636bb-c76a-4895-8af6-6801f45e4e5b\") " pod="openstack/keystone-c4f8d8d9b-p5pqq" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.984666 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6d636bb-c76a-4895-8af6-6801f45e4e5b-scripts\") pod \"keystone-c4f8d8d9b-p5pqq\" (UID: \"c6d636bb-c76a-4895-8af6-6801f45e4e5b\") " pod="openstack/keystone-c4f8d8d9b-p5pqq" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.984691 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6d636bb-c76a-4895-8af6-6801f45e4e5b-config-data\") pod \"keystone-c4f8d8d9b-p5pqq\" (UID: \"c6d636bb-c76a-4895-8af6-6801f45e4e5b\") " pod="openstack/keystone-c4f8d8d9b-p5pqq" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.984731 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fx8v\" (UniqueName: \"kubernetes.io/projected/c6d636bb-c76a-4895-8af6-6801f45e4e5b-kube-api-access-8fx8v\") pod \"keystone-c4f8d8d9b-p5pqq\" (UID: \"c6d636bb-c76a-4895-8af6-6801f45e4e5b\") " pod="openstack/keystone-c4f8d8d9b-p5pqq" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.984747 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6d636bb-c76a-4895-8af6-6801f45e4e5b-combined-ca-bundle\") pod \"keystone-c4f8d8d9b-p5pqq\" (UID: \"c6d636bb-c76a-4895-8af6-6801f45e4e5b\") " pod="openstack/keystone-c4f8d8d9b-p5pqq" Dec 06 16:03:32 crc kubenswrapper[4813]: I1206 16:03:32.990382 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c6d636bb-c76a-4895-8af6-6801f45e4e5b-fernet-keys\") pod \"keystone-c4f8d8d9b-p5pqq\" (UID: \"c6d636bb-c76a-4895-8af6-6801f45e4e5b\") " pod="openstack/keystone-c4f8d8d9b-p5pqq" Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.084647 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-64449f75dd-k89r7" event={"ID":"322df0dd-86e6-466b-8afa-e4e52e889827","Type":"ContainerStarted","Data":"59817fb2ddd3d0ceea3f1c424f39c090db148fe2df3e2a62f0634789515c7268"} Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.087023 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-ffc86b7c-26pxs"] Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.099460 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-66f946f97d-gh5xc" event={"ID":"78cfb04c-4099-4211-a82e-47a304251ab5","Type":"ContainerStarted","Data":"1a239a6a5b6ee4d213a60af0e3d2c40cd7a6db930283f98e9065fa779c7d3a86"} Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.100603 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6d636bb-c76a-4895-8af6-6801f45e4e5b-internal-tls-certs\") pod \"keystone-c4f8d8d9b-p5pqq\" (UID: \"c6d636bb-c76a-4895-8af6-6801f45e4e5b\") " pod="openstack/keystone-c4f8d8d9b-p5pqq" Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.100742 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6d636bb-c76a-4895-8af6-6801f45e4e5b-scripts\") pod \"keystone-c4f8d8d9b-p5pqq\" (UID: \"c6d636bb-c76a-4895-8af6-6801f45e4e5b\") " pod="openstack/keystone-c4f8d8d9b-p5pqq" Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.101679 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6d636bb-c76a-4895-8af6-6801f45e4e5b-config-data\") pod \"keystone-c4f8d8d9b-p5pqq\" (UID: \"c6d636bb-c76a-4895-8af6-6801f45e4e5b\") " pod="openstack/keystone-c4f8d8d9b-p5pqq" Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.101790 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fx8v\" (UniqueName: \"kubernetes.io/projected/c6d636bb-c76a-4895-8af6-6801f45e4e5b-kube-api-access-8fx8v\") pod \"keystone-c4f8d8d9b-p5pqq\" (UID: \"c6d636bb-c76a-4895-8af6-6801f45e4e5b\") " pod="openstack/keystone-c4f8d8d9b-p5pqq" Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.101855 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6d636bb-c76a-4895-8af6-6801f45e4e5b-combined-ca-bundle\") pod \"keystone-c4f8d8d9b-p5pqq\" (UID: \"c6d636bb-c76a-4895-8af6-6801f45e4e5b\") " pod="openstack/keystone-c4f8d8d9b-p5pqq" Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.101920 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c6d636bb-c76a-4895-8af6-6801f45e4e5b-fernet-keys\") pod \"keystone-c4f8d8d9b-p5pqq\" (UID: \"c6d636bb-c76a-4895-8af6-6801f45e4e5b\") " pod="openstack/keystone-c4f8d8d9b-p5pqq" Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.101998 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6d636bb-c76a-4895-8af6-6801f45e4e5b-public-tls-certs\") pod \"keystone-c4f8d8d9b-p5pqq\" (UID: \"c6d636bb-c76a-4895-8af6-6801f45e4e5b\") " pod="openstack/keystone-c4f8d8d9b-p5pqq" Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.102116 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c6d636bb-c76a-4895-8af6-6801f45e4e5b-credential-keys\") pod \"keystone-c4f8d8d9b-p5pqq\" (UID: \"c6d636bb-c76a-4895-8af6-6801f45e4e5b\") " pod="openstack/keystone-c4f8d8d9b-p5pqq" Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.107972 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6d636bb-c76a-4895-8af6-6801f45e4e5b-scripts\") pod \"keystone-c4f8d8d9b-p5pqq\" (UID: \"c6d636bb-c76a-4895-8af6-6801f45e4e5b\") " pod="openstack/keystone-c4f8d8d9b-p5pqq" Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.108491 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c6d636bb-c76a-4895-8af6-6801f45e4e5b-fernet-keys\") pod \"keystone-c4f8d8d9b-p5pqq\" (UID: \"c6d636bb-c76a-4895-8af6-6801f45e4e5b\") " pod="openstack/keystone-c4f8d8d9b-p5pqq" Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.111584 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c6d636bb-c76a-4895-8af6-6801f45e4e5b-credential-keys\") pod \"keystone-c4f8d8d9b-p5pqq\" (UID: \"c6d636bb-c76a-4895-8af6-6801f45e4e5b\") " pod="openstack/keystone-c4f8d8d9b-p5pqq" Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.111885 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6d636bb-c76a-4895-8af6-6801f45e4e5b-config-data\") pod \"keystone-c4f8d8d9b-p5pqq\" (UID: \"c6d636bb-c76a-4895-8af6-6801f45e4e5b\") " pod="openstack/keystone-c4f8d8d9b-p5pqq" Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.112409 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6d636bb-c76a-4895-8af6-6801f45e4e5b-internal-tls-certs\") pod \"keystone-c4f8d8d9b-p5pqq\" (UID: \"c6d636bb-c76a-4895-8af6-6801f45e4e5b\") " pod="openstack/keystone-c4f8d8d9b-p5pqq" Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.120087 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6d636bb-c76a-4895-8af6-6801f45e4e5b-public-tls-certs\") pod \"keystone-c4f8d8d9b-p5pqq\" (UID: \"c6d636bb-c76a-4895-8af6-6801f45e4e5b\") " pod="openstack/keystone-c4f8d8d9b-p5pqq" Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.124678 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fx8v\" (UniqueName: \"kubernetes.io/projected/c6d636bb-c76a-4895-8af6-6801f45e4e5b-kube-api-access-8fx8v\") pod \"keystone-c4f8d8d9b-p5pqq\" (UID: \"c6d636bb-c76a-4895-8af6-6801f45e4e5b\") " pod="openstack/keystone-c4f8d8d9b-p5pqq" Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.130232 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-98dddfffb-l698f" event={"ID":"e913dfcd-dfe3-4c2c-b72a-8393295c95fc","Type":"ContainerStarted","Data":"b0b03946c6de0dc3c9be001798e1235034503d15c4fadb6e377ca49014f72957"} Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.132910 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6d636bb-c76a-4895-8af6-6801f45e4e5b-combined-ca-bundle\") pod \"keystone-c4f8d8d9b-p5pqq\" (UID: \"c6d636bb-c76a-4895-8af6-6801f45e4e5b\") " pod="openstack/keystone-c4f8d8d9b-p5pqq" Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.167442 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" event={"ID":"90124024-daf7-4241-a787-2840359fe074","Type":"ContainerDied","Data":"59262c1305d593a6033d2261739d33f8ab5c31d54f9b90e00777e37328674810"} Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.167484 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-qdtmw" Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.167490 4813 scope.go:117] "RemoveContainer" containerID="b4833ba9d04cff0f35659eeb3b203e748fe5c3214b9fed117882cb9e9972b2b8" Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.180021 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58d4d548d-6bl8w" event={"ID":"f883aaa3-f010-4c52-bc12-1591fb0ddb4d","Type":"ContainerStarted","Data":"297489db702f144b7e8d1f1e9200a74e4aa43293ea6acc847fe523ace3575f60"} Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.198663 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" event={"ID":"264bf858-a126-4005-9c98-0515834c4956","Type":"ContainerStarted","Data":"6cacbbb17ecd4d71a0f31e137350bebb3c7ae7742fcc9fa5a2350837b8e034d1"} Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.222489 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-qdtmw"] Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.223546 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.223574 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.224338 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f539e8ab-300c-426b-98e7-f9e87a6787b5","Type":"ContainerStarted","Data":"cf02350d626d2d9a7dd8311ef157f7dd95a5cbcbf89c6fccc653e5627ca1f6cd"} Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.224393 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.224402 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.233451 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-qdtmw"] Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.238799 4813 scope.go:117] "RemoveContainer" containerID="0a7a2ceb51aff58c096e8545422f4025de69e0d87601130d3445d785e5a918ce" Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.244700 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c4f8d8d9b-p5pqq" Dec 06 16:03:33 crc kubenswrapper[4813]: I1206 16:03:33.989565 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-c4f8d8d9b-p5pqq"] Dec 06 16:03:34 crc kubenswrapper[4813]: I1206 16:03:34.243792 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58d4d548d-6bl8w" event={"ID":"f883aaa3-f010-4c52-bc12-1591fb0ddb4d","Type":"ContainerStarted","Data":"5ebc20c08927e8a4a7bc2fccd6c5329175ecbeb6c9e4f31e98c656d01a8a69e7"} Dec 06 16:03:34 crc kubenswrapper[4813]: I1206 16:03:34.243841 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58d4d548d-6bl8w" event={"ID":"f883aaa3-f010-4c52-bc12-1591fb0ddb4d","Type":"ContainerStarted","Data":"31c54f8945ba2cb3728203e5d282b17de808bdb5f02f6d468105b7a7bbdd419a"} Dec 06 16:03:34 crc kubenswrapper[4813]: I1206 16:03:34.244332 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-58d4d548d-6bl8w" Dec 06 16:03:34 crc kubenswrapper[4813]: I1206 16:03:34.244446 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-58d4d548d-6bl8w" Dec 06 16:03:34 crc kubenswrapper[4813]: I1206 16:03:34.257770 4813 generic.go:334] "Generic (PLEG): container finished" podID="264bf858-a126-4005-9c98-0515834c4956" containerID="08467d4834e76df9230e68f5f05769d43c566c623bc6263be28d1c7a26156095" exitCode=0 Dec 06 16:03:34 crc kubenswrapper[4813]: I1206 16:03:34.257825 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" event={"ID":"264bf858-a126-4005-9c98-0515834c4956","Type":"ContainerDied","Data":"08467d4834e76df9230e68f5f05769d43c566c623bc6263be28d1c7a26156095"} Dec 06 16:03:34 crc kubenswrapper[4813]: I1206 16:03:34.268310 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-ffc86b7c-26pxs" event={"ID":"cf88c989-e78b-456e-ba32-19d8af98a1d0","Type":"ContainerStarted","Data":"1875be6af023d1f2e19c9b4002081348de03659e5e0e82152713c48fcce8a82c"} Dec 06 16:03:34 crc kubenswrapper[4813]: I1206 16:03:34.271544 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-58d4d548d-6bl8w" podStartSLOduration=5.271522727 podStartE2EDuration="5.271522727s" podCreationTimestamp="2025-12-06 16:03:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:03:34.260440966 +0000 UTC m=+1054.151320542" watchObservedRunningTime="2025-12-06 16:03:34.271522727 +0000 UTC m=+1054.162402313" Dec 06 16:03:34 crc kubenswrapper[4813]: I1206 16:03:34.278244 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c4f8d8d9b-p5pqq" event={"ID":"c6d636bb-c76a-4895-8af6-6801f45e4e5b","Type":"ContainerStarted","Data":"d8027c2d424d0fb45ae0746cd070082704a96c2a9db8db24bd05e3c2cd591707"} Dec 06 16:03:34 crc kubenswrapper[4813]: I1206 16:03:34.288461 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-66f946f97d-gh5xc" event={"ID":"78cfb04c-4099-4211-a82e-47a304251ab5","Type":"ContainerStarted","Data":"acc4f465e1678852c5cec3f2ce1dda35b19bb8e717db2826e371bd238aef5f09"} Dec 06 16:03:34 crc kubenswrapper[4813]: I1206 16:03:34.288505 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-66f946f97d-gh5xc" event={"ID":"78cfb04c-4099-4211-a82e-47a304251ab5","Type":"ContainerStarted","Data":"28a8a23ceae65c8b9be7f360a346a712fe0b76eaa4ec1db1bad715488651fd35"} Dec 06 16:03:34 crc kubenswrapper[4813]: I1206 16:03:34.289514 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-66f946f97d-gh5xc" Dec 06 16:03:34 crc kubenswrapper[4813]: I1206 16:03:34.289543 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-66f946f97d-gh5xc" Dec 06 16:03:34 crc kubenswrapper[4813]: I1206 16:03:34.304048 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-98dddfffb-l698f" event={"ID":"e913dfcd-dfe3-4c2c-b72a-8393295c95fc","Type":"ContainerStarted","Data":"6adb826e2c36b4ad9be9167482d865f172767f06e6222d56d8a23a6aad91766d"} Dec 06 16:03:34 crc kubenswrapper[4813]: I1206 16:03:34.304095 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-98dddfffb-l698f" event={"ID":"e913dfcd-dfe3-4c2c-b72a-8393295c95fc","Type":"ContainerStarted","Data":"96a8bb31fe9274580e6e22d95ce0ea04459044399f6ba35dd973d0e1cbc559fd"} Dec 06 16:03:34 crc kubenswrapper[4813]: I1206 16:03:34.308767 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-98dddfffb-l698f" Dec 06 16:03:34 crc kubenswrapper[4813]: I1206 16:03:34.308978 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-98dddfffb-l698f" Dec 06 16:03:34 crc kubenswrapper[4813]: I1206 16:03:34.318820 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-66f946f97d-gh5xc" podStartSLOduration=9.318801629 podStartE2EDuration="9.318801629s" podCreationTimestamp="2025-12-06 16:03:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:03:34.315479112 +0000 UTC m=+1054.206358688" watchObservedRunningTime="2025-12-06 16:03:34.318801629 +0000 UTC m=+1054.209681205" Dec 06 16:03:34 crc kubenswrapper[4813]: I1206 16:03:34.342932 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-lvw7n" event={"ID":"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32","Type":"ContainerStarted","Data":"e8a3382caedf6fb750d7213c7c4ffe6af9b5c1791f066d46a29aff583c43e72a"} Dec 06 16:03:34 crc kubenswrapper[4813]: I1206 16:03:34.347231 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-98dddfffb-l698f" podStartSLOduration=10.347208925 podStartE2EDuration="10.347208925s" podCreationTimestamp="2025-12-06 16:03:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:03:34.341549466 +0000 UTC m=+1054.232429042" watchObservedRunningTime="2025-12-06 16:03:34.347208925 +0000 UTC m=+1054.238088491" Dec 06 16:03:34 crc kubenswrapper[4813]: I1206 16:03:34.369845 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-lvw7n" podStartSLOduration=4.422182889 podStartE2EDuration="57.369826539s" podCreationTimestamp="2025-12-06 16:02:37 +0000 UTC" firstStartedPulling="2025-12-06 16:02:39.05939171 +0000 UTC m=+998.950271286" lastFinishedPulling="2025-12-06 16:03:32.00703536 +0000 UTC m=+1051.897914936" observedRunningTime="2025-12-06 16:03:34.358478661 +0000 UTC m=+1054.249358237" watchObservedRunningTime="2025-12-06 16:03:34.369826539 +0000 UTC m=+1054.260706105" Dec 06 16:03:34 crc kubenswrapper[4813]: I1206 16:03:34.498193 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90124024-daf7-4241-a787-2840359fe074" path="/var/lib/kubelet/pods/90124024-daf7-4241-a787-2840359fe074/volumes" Dec 06 16:03:35 crc kubenswrapper[4813]: I1206 16:03:35.376374 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" event={"ID":"264bf858-a126-4005-9c98-0515834c4956","Type":"ContainerStarted","Data":"399760fcc15809e2f6b6f950ca184113e25ea3d797744dab292c4dc1e9da448d"} Dec 06 16:03:35 crc kubenswrapper[4813]: I1206 16:03:35.377456 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" Dec 06 16:03:35 crc kubenswrapper[4813]: I1206 16:03:35.403995 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" podStartSLOduration=10.403980356 podStartE2EDuration="10.403980356s" podCreationTimestamp="2025-12-06 16:03:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:03:35.402387914 +0000 UTC m=+1055.293267490" watchObservedRunningTime="2025-12-06 16:03:35.403980356 +0000 UTC m=+1055.294859932" Dec 06 16:03:35 crc kubenswrapper[4813]: I1206 16:03:35.405242 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c4f8d8d9b-p5pqq" event={"ID":"c6d636bb-c76a-4895-8af6-6801f45e4e5b","Type":"ContainerStarted","Data":"a2031d1efcd49bbc6d03bfb0a93c34b51066bc5967ff0e91c71e549ab4dc6a8d"} Dec 06 16:03:35 crc kubenswrapper[4813]: I1206 16:03:35.406154 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-c4f8d8d9b-p5pqq" Dec 06 16:03:35 crc kubenswrapper[4813]: I1206 16:03:35.438684 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-c4f8d8d9b-p5pqq" podStartSLOduration=3.438664447 podStartE2EDuration="3.438664447s" podCreationTimestamp="2025-12-06 16:03:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:03:35.433032839 +0000 UTC m=+1055.323912425" watchObservedRunningTime="2025-12-06 16:03:35.438664447 +0000 UTC m=+1055.329544023" Dec 06 16:03:36 crc kubenswrapper[4813]: I1206 16:03:36.232764 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-bbc668d58-h8w67" podUID="4bdc8329-162b-4592-b98e-ec7eda7f2ce1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Dec 06 16:03:36 crc kubenswrapper[4813]: I1206 16:03:36.345784 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7fd6f7946b-kmm96" podUID="28238998-5d44-4ae9-8d1f-93d56ff18152" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Dec 06 16:03:36 crc kubenswrapper[4813]: I1206 16:03:36.539798 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 06 16:03:36 crc kubenswrapper[4813]: I1206 16:03:36.539900 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 16:03:36 crc kubenswrapper[4813]: I1206 16:03:36.547010 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 06 16:03:36 crc kubenswrapper[4813]: I1206 16:03:36.760316 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 06 16:03:36 crc kubenswrapper[4813]: I1206 16:03:36.760421 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 16:03:36 crc kubenswrapper[4813]: I1206 16:03:36.766105 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 06 16:03:38 crc kubenswrapper[4813]: I1206 16:03:38.447935 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-64449f75dd-k89r7" event={"ID":"322df0dd-86e6-466b-8afa-e4e52e889827","Type":"ContainerStarted","Data":"cb871b54b25f298d2092e7d8d7b6a4ec3b34e0dab689bb45b0dcca08c2ba23a9"} Dec 06 16:03:38 crc kubenswrapper[4813]: I1206 16:03:38.448196 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-64449f75dd-k89r7" event={"ID":"322df0dd-86e6-466b-8afa-e4e52e889827","Type":"ContainerStarted","Data":"31af8b035f2ba01f316eb4e0b8b9362914b9daf8b8b759d06a41fd3944b7a00f"} Dec 06 16:03:38 crc kubenswrapper[4813]: I1206 16:03:38.454363 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-ffc86b7c-26pxs" event={"ID":"cf88c989-e78b-456e-ba32-19d8af98a1d0","Type":"ContainerStarted","Data":"71ff27a20cededd2568f59ba010a2b46a852f9a87cfac2fc6e913fe2814c82b1"} Dec 06 16:03:38 crc kubenswrapper[4813]: I1206 16:03:38.454401 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-ffc86b7c-26pxs" event={"ID":"cf88c989-e78b-456e-ba32-19d8af98a1d0","Type":"ContainerStarted","Data":"59f2e5c67ca44e03998f180f3b6aa0881c5c05cff0f67db21bd4517b7f6bdcda"} Dec 06 16:03:38 crc kubenswrapper[4813]: I1206 16:03:38.474320 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-64449f75dd-k89r7" podStartSLOduration=8.50424734 podStartE2EDuration="13.474304762s" podCreationTimestamp="2025-12-06 16:03:25 +0000 UTC" firstStartedPulling="2025-12-06 16:03:32.716146897 +0000 UTC m=+1052.607026473" lastFinishedPulling="2025-12-06 16:03:37.686204319 +0000 UTC m=+1057.577083895" observedRunningTime="2025-12-06 16:03:38.47233623 +0000 UTC m=+1058.363215806" watchObservedRunningTime="2025-12-06 16:03:38.474304762 +0000 UTC m=+1058.365184338" Dec 06 16:03:38 crc kubenswrapper[4813]: I1206 16:03:38.490567 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-ffc86b7c-26pxs" podStartSLOduration=8.972216784 podStartE2EDuration="13.490552029s" podCreationTimestamp="2025-12-06 16:03:25 +0000 UTC" firstStartedPulling="2025-12-06 16:03:33.170848012 +0000 UTC m=+1053.061727588" lastFinishedPulling="2025-12-06 16:03:37.689183257 +0000 UTC m=+1057.580062833" observedRunningTime="2025-12-06 16:03:38.489003708 +0000 UTC m=+1058.379883284" watchObservedRunningTime="2025-12-06 16:03:38.490552029 +0000 UTC m=+1058.381431605" Dec 06 16:03:40 crc kubenswrapper[4813]: I1206 16:03:40.644894 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" Dec 06 16:03:40 crc kubenswrapper[4813]: I1206 16:03:40.747037 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-rrfrw"] Dec 06 16:03:40 crc kubenswrapper[4813]: I1206 16:03:40.747246 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" podUID="efca4ca3-8021-4417-83cc-1b26f13f97bd" containerName="dnsmasq-dns" containerID="cri-o://2b0d09d239154fc3944e334774798b0da04c92d93b83aee3bbb341b976d5155c" gracePeriod=10 Dec 06 16:03:41 crc kubenswrapper[4813]: I1206 16:03:41.481101 4813 generic.go:334] "Generic (PLEG): container finished" podID="efca4ca3-8021-4417-83cc-1b26f13f97bd" containerID="2b0d09d239154fc3944e334774798b0da04c92d93b83aee3bbb341b976d5155c" exitCode=0 Dec 06 16:03:41 crc kubenswrapper[4813]: I1206 16:03:41.481323 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" event={"ID":"efca4ca3-8021-4417-83cc-1b26f13f97bd","Type":"ContainerDied","Data":"2b0d09d239154fc3944e334774798b0da04c92d93b83aee3bbb341b976d5155c"} Dec 06 16:03:41 crc kubenswrapper[4813]: I1206 16:03:41.562195 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-58d4d548d-6bl8w" Dec 06 16:03:42 crc kubenswrapper[4813]: I1206 16:03:42.264674 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-58d4d548d-6bl8w" Dec 06 16:03:42 crc kubenswrapper[4813]: I1206 16:03:42.396453 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-66f946f97d-gh5xc"] Dec 06 16:03:42 crc kubenswrapper[4813]: I1206 16:03:42.396695 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-66f946f97d-gh5xc" podUID="78cfb04c-4099-4211-a82e-47a304251ab5" containerName="barbican-api-log" containerID="cri-o://28a8a23ceae65c8b9be7f360a346a712fe0b76eaa4ec1db1bad715488651fd35" gracePeriod=30 Dec 06 16:03:42 crc kubenswrapper[4813]: I1206 16:03:42.396713 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-66f946f97d-gh5xc" podUID="78cfb04c-4099-4211-a82e-47a304251ab5" containerName="barbican-api" containerID="cri-o://acc4f465e1678852c5cec3f2ce1dda35b19bb8e717db2826e371bd238aef5f09" gracePeriod=30 Dec 06 16:03:42 crc kubenswrapper[4813]: I1206 16:03:42.412674 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-66f946f97d-gh5xc" podUID="78cfb04c-4099-4211-a82e-47a304251ab5" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": EOF" Dec 06 16:03:42 crc kubenswrapper[4813]: I1206 16:03:42.413017 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-66f946f97d-gh5xc" podUID="78cfb04c-4099-4211-a82e-47a304251ab5" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": EOF" Dec 06 16:03:42 crc kubenswrapper[4813]: I1206 16:03:42.413223 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-66f946f97d-gh5xc" podUID="78cfb04c-4099-4211-a82e-47a304251ab5" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": EOF" Dec 06 16:03:42 crc kubenswrapper[4813]: I1206 16:03:42.413467 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-66f946f97d-gh5xc" podUID="78cfb04c-4099-4211-a82e-47a304251ab5" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": EOF" Dec 06 16:03:42 crc kubenswrapper[4813]: I1206 16:03:42.527934 4813 generic.go:334] "Generic (PLEG): container finished" podID="8cf4a1cf-ef96-4e50-946c-cd0c5eccff32" containerID="e8a3382caedf6fb750d7213c7c4ffe6af9b5c1791f066d46a29aff583c43e72a" exitCode=0 Dec 06 16:03:42 crc kubenswrapper[4813]: I1206 16:03:42.528873 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-lvw7n" event={"ID":"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32","Type":"ContainerDied","Data":"e8a3382caedf6fb750d7213c7c4ffe6af9b5c1791f066d46a29aff583c43e72a"} Dec 06 16:03:43 crc kubenswrapper[4813]: I1206 16:03:43.537849 4813 generic.go:334] "Generic (PLEG): container finished" podID="78cfb04c-4099-4211-a82e-47a304251ab5" containerID="28a8a23ceae65c8b9be7f360a346a712fe0b76eaa4ec1db1bad715488651fd35" exitCode=143 Dec 06 16:03:43 crc kubenswrapper[4813]: I1206 16:03:43.537931 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-66f946f97d-gh5xc" event={"ID":"78cfb04c-4099-4211-a82e-47a304251ab5","Type":"ContainerDied","Data":"28a8a23ceae65c8b9be7f360a346a712fe0b76eaa4ec1db1bad715488651fd35"} Dec 06 16:03:43 crc kubenswrapper[4813]: I1206 16:03:43.630991 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" podUID="efca4ca3-8021-4417-83cc-1b26f13f97bd" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.145:5353: connect: connection refused" Dec 06 16:03:45 crc kubenswrapper[4813]: I1206 16:03:45.555253 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7c8fc98b5b-dwkdc" Dec 06 16:03:46 crc kubenswrapper[4813]: I1206 16:03:46.232046 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-bbc668d58-h8w67" podUID="4bdc8329-162b-4592-b98e-ec7eda7f2ce1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Dec 06 16:03:46 crc kubenswrapper[4813]: I1206 16:03:46.344768 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7fd6f7946b-kmm96" podUID="28238998-5d44-4ae9-8d1f-93d56ff18152" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Dec 06 16:03:47 crc kubenswrapper[4813]: I1206 16:03:47.083294 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-lvw7n" Dec 06 16:03:47 crc kubenswrapper[4813]: I1206 16:03:47.250123 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-combined-ca-bundle\") pod \"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32\" (UID: \"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32\") " Dec 06 16:03:47 crc kubenswrapper[4813]: I1206 16:03:47.252407 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-scripts\") pod \"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32\" (UID: \"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32\") " Dec 06 16:03:47 crc kubenswrapper[4813]: I1206 16:03:47.252495 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpz7c\" (UniqueName: \"kubernetes.io/projected/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-kube-api-access-cpz7c\") pod \"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32\" (UID: \"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32\") " Dec 06 16:03:47 crc kubenswrapper[4813]: I1206 16:03:47.252567 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-db-sync-config-data\") pod \"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32\" (UID: \"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32\") " Dec 06 16:03:47 crc kubenswrapper[4813]: I1206 16:03:47.252612 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-etc-machine-id\") pod \"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32\" (UID: \"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32\") " Dec 06 16:03:47 crc kubenswrapper[4813]: I1206 16:03:47.252870 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-config-data\") pod \"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32\" (UID: \"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32\") " Dec 06 16:03:47 crc kubenswrapper[4813]: I1206 16:03:47.253638 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "8cf4a1cf-ef96-4e50-946c-cd0c5eccff32" (UID: "8cf4a1cf-ef96-4e50-946c-cd0c5eccff32"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 16:03:47 crc kubenswrapper[4813]: I1206 16:03:47.253929 4813 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:47 crc kubenswrapper[4813]: I1206 16:03:47.259431 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-kube-api-access-cpz7c" (OuterVolumeSpecName: "kube-api-access-cpz7c") pod "8cf4a1cf-ef96-4e50-946c-cd0c5eccff32" (UID: "8cf4a1cf-ef96-4e50-946c-cd0c5eccff32"). InnerVolumeSpecName "kube-api-access-cpz7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:03:47 crc kubenswrapper[4813]: I1206 16:03:47.290870 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-scripts" (OuterVolumeSpecName: "scripts") pod "8cf4a1cf-ef96-4e50-946c-cd0c5eccff32" (UID: "8cf4a1cf-ef96-4e50-946c-cd0c5eccff32"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:47 crc kubenswrapper[4813]: I1206 16:03:47.290926 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "8cf4a1cf-ef96-4e50-946c-cd0c5eccff32" (UID: "8cf4a1cf-ef96-4e50-946c-cd0c5eccff32"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:47 crc kubenswrapper[4813]: I1206 16:03:47.297370 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8cf4a1cf-ef96-4e50-946c-cd0c5eccff32" (UID: "8cf4a1cf-ef96-4e50-946c-cd0c5eccff32"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:47 crc kubenswrapper[4813]: I1206 16:03:47.322602 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-config-data" (OuterVolumeSpecName: "config-data") pod "8cf4a1cf-ef96-4e50-946c-cd0c5eccff32" (UID: "8cf4a1cf-ef96-4e50-946c-cd0c5eccff32"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:47 crc kubenswrapper[4813]: I1206 16:03:47.355791 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:47 crc kubenswrapper[4813]: I1206 16:03:47.355828 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:47 crc kubenswrapper[4813]: I1206 16:03:47.355841 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpz7c\" (UniqueName: \"kubernetes.io/projected/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-kube-api-access-cpz7c\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:47 crc kubenswrapper[4813]: I1206 16:03:47.355855 4813 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:47 crc kubenswrapper[4813]: I1206 16:03:47.355866 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:47 crc kubenswrapper[4813]: I1206 16:03:47.496456 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-66f946f97d-gh5xc" podUID="78cfb04c-4099-4211-a82e-47a304251ab5" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 16:03:47 crc kubenswrapper[4813]: I1206 16:03:47.496732 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-66f946f97d-gh5xc" podUID="78cfb04c-4099-4211-a82e-47a304251ab5" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 16:03:47 crc kubenswrapper[4813]: I1206 16:03:47.647057 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-lvw7n" event={"ID":"8cf4a1cf-ef96-4e50-946c-cd0c5eccff32","Type":"ContainerDied","Data":"22d0a046133d14bb5dc9621fe51f7f033ea8f91fedd722681ab075a659d6a00c"} Dec 06 16:03:47 crc kubenswrapper[4813]: I1206 16:03:47.647109 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-lvw7n" Dec 06 16:03:47 crc kubenswrapper[4813]: I1206 16:03:47.647113 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22d0a046133d14bb5dc9621fe51f7f033ea8f91fedd722681ab075a659d6a00c" Dec 06 16:03:47 crc kubenswrapper[4813]: I1206 16:03:47.829876 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-66f946f97d-gh5xc" podUID="78cfb04c-4099-4211-a82e-47a304251ab5" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:41582->10.217.0.160:9311: read: connection reset by peer" Dec 06 16:03:47 crc kubenswrapper[4813]: I1206 16:03:47.830334 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-66f946f97d-gh5xc" podUID="78cfb04c-4099-4211-a82e-47a304251ab5" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:41594->10.217.0.160:9311: read: connection reset by peer" Dec 06 16:03:47 crc kubenswrapper[4813]: I1206 16:03:47.935348 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" Dec 06 16:03:48 crc kubenswrapper[4813]: E1206 16:03:48.052949 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="f539e8ab-300c-426b-98e7-f9e87a6787b5" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.066076 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/efca4ca3-8021-4417-83cc-1b26f13f97bd-ovsdbserver-nb\") pod \"efca4ca3-8021-4417-83cc-1b26f13f97bd\" (UID: \"efca4ca3-8021-4417-83cc-1b26f13f97bd\") " Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.066114 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/efca4ca3-8021-4417-83cc-1b26f13f97bd-dns-svc\") pod \"efca4ca3-8021-4417-83cc-1b26f13f97bd\" (UID: \"efca4ca3-8021-4417-83cc-1b26f13f97bd\") " Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.066187 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/efca4ca3-8021-4417-83cc-1b26f13f97bd-ovsdbserver-sb\") pod \"efca4ca3-8021-4417-83cc-1b26f13f97bd\" (UID: \"efca4ca3-8021-4417-83cc-1b26f13f97bd\") " Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.066282 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/efca4ca3-8021-4417-83cc-1b26f13f97bd-dns-swift-storage-0\") pod \"efca4ca3-8021-4417-83cc-1b26f13f97bd\" (UID: \"efca4ca3-8021-4417-83cc-1b26f13f97bd\") " Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.066327 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efca4ca3-8021-4417-83cc-1b26f13f97bd-config\") pod \"efca4ca3-8021-4417-83cc-1b26f13f97bd\" (UID: \"efca4ca3-8021-4417-83cc-1b26f13f97bd\") " Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.066374 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxpln\" (UniqueName: \"kubernetes.io/projected/efca4ca3-8021-4417-83cc-1b26f13f97bd-kube-api-access-wxpln\") pod \"efca4ca3-8021-4417-83cc-1b26f13f97bd\" (UID: \"efca4ca3-8021-4417-83cc-1b26f13f97bd\") " Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.070214 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efca4ca3-8021-4417-83cc-1b26f13f97bd-kube-api-access-wxpln" (OuterVolumeSpecName: "kube-api-access-wxpln") pod "efca4ca3-8021-4417-83cc-1b26f13f97bd" (UID: "efca4ca3-8021-4417-83cc-1b26f13f97bd"). InnerVolumeSpecName "kube-api-access-wxpln". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.133375 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efca4ca3-8021-4417-83cc-1b26f13f97bd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "efca4ca3-8021-4417-83cc-1b26f13f97bd" (UID: "efca4ca3-8021-4417-83cc-1b26f13f97bd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.143897 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efca4ca3-8021-4417-83cc-1b26f13f97bd-config" (OuterVolumeSpecName: "config") pod "efca4ca3-8021-4417-83cc-1b26f13f97bd" (UID: "efca4ca3-8021-4417-83cc-1b26f13f97bd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.154869 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efca4ca3-8021-4417-83cc-1b26f13f97bd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "efca4ca3-8021-4417-83cc-1b26f13f97bd" (UID: "efca4ca3-8021-4417-83cc-1b26f13f97bd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.161740 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efca4ca3-8021-4417-83cc-1b26f13f97bd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "efca4ca3-8021-4417-83cc-1b26f13f97bd" (UID: "efca4ca3-8021-4417-83cc-1b26f13f97bd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.169866 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/efca4ca3-8021-4417-83cc-1b26f13f97bd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.169913 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/efca4ca3-8021-4417-83cc-1b26f13f97bd-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.169925 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/efca4ca3-8021-4417-83cc-1b26f13f97bd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.169952 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efca4ca3-8021-4417-83cc-1b26f13f97bd-config\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.169962 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxpln\" (UniqueName: \"kubernetes.io/projected/efca4ca3-8021-4417-83cc-1b26f13f97bd-kube-api-access-wxpln\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.173491 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efca4ca3-8021-4417-83cc-1b26f13f97bd-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "efca4ca3-8021-4417-83cc-1b26f13f97bd" (UID: "efca4ca3-8021-4417-83cc-1b26f13f97bd"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.244015 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-66f946f97d-gh5xc" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.271290 4813 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/efca4ca3-8021-4417-83cc-1b26f13f97bd-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.372034 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78cfb04c-4099-4211-a82e-47a304251ab5-config-data-custom\") pod \"78cfb04c-4099-4211-a82e-47a304251ab5\" (UID: \"78cfb04c-4099-4211-a82e-47a304251ab5\") " Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.372116 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78cfb04c-4099-4211-a82e-47a304251ab5-config-data\") pod \"78cfb04c-4099-4211-a82e-47a304251ab5\" (UID: \"78cfb04c-4099-4211-a82e-47a304251ab5\") " Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.372169 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78cfb04c-4099-4211-a82e-47a304251ab5-combined-ca-bundle\") pod \"78cfb04c-4099-4211-a82e-47a304251ab5\" (UID: \"78cfb04c-4099-4211-a82e-47a304251ab5\") " Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.372251 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bwn7\" (UniqueName: \"kubernetes.io/projected/78cfb04c-4099-4211-a82e-47a304251ab5-kube-api-access-6bwn7\") pod \"78cfb04c-4099-4211-a82e-47a304251ab5\" (UID: \"78cfb04c-4099-4211-a82e-47a304251ab5\") " Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.372365 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78cfb04c-4099-4211-a82e-47a304251ab5-logs\") pod \"78cfb04c-4099-4211-a82e-47a304251ab5\" (UID: \"78cfb04c-4099-4211-a82e-47a304251ab5\") " Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.372987 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78cfb04c-4099-4211-a82e-47a304251ab5-logs" (OuterVolumeSpecName: "logs") pod "78cfb04c-4099-4211-a82e-47a304251ab5" (UID: "78cfb04c-4099-4211-a82e-47a304251ab5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.382737 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78cfb04c-4099-4211-a82e-47a304251ab5-kube-api-access-6bwn7" (OuterVolumeSpecName: "kube-api-access-6bwn7") pod "78cfb04c-4099-4211-a82e-47a304251ab5" (UID: "78cfb04c-4099-4211-a82e-47a304251ab5"). InnerVolumeSpecName "kube-api-access-6bwn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.382866 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78cfb04c-4099-4211-a82e-47a304251ab5-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "78cfb04c-4099-4211-a82e-47a304251ab5" (UID: "78cfb04c-4099-4211-a82e-47a304251ab5"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.475953 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 16:03:48 crc kubenswrapper[4813]: E1206 16:03:48.476294 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78cfb04c-4099-4211-a82e-47a304251ab5" containerName="barbican-api-log" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.476313 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="78cfb04c-4099-4211-a82e-47a304251ab5" containerName="barbican-api-log" Dec 06 16:03:48 crc kubenswrapper[4813]: E1206 16:03:48.476325 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efca4ca3-8021-4417-83cc-1b26f13f97bd" containerName="init" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.476332 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="efca4ca3-8021-4417-83cc-1b26f13f97bd" containerName="init" Dec 06 16:03:48 crc kubenswrapper[4813]: E1206 16:03:48.476340 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cf4a1cf-ef96-4e50-946c-cd0c5eccff32" containerName="cinder-db-sync" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.476346 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cf4a1cf-ef96-4e50-946c-cd0c5eccff32" containerName="cinder-db-sync" Dec 06 16:03:48 crc kubenswrapper[4813]: E1206 16:03:48.476361 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efca4ca3-8021-4417-83cc-1b26f13f97bd" containerName="dnsmasq-dns" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.476367 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="efca4ca3-8021-4417-83cc-1b26f13f97bd" containerName="dnsmasq-dns" Dec 06 16:03:48 crc kubenswrapper[4813]: E1206 16:03:48.476401 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78cfb04c-4099-4211-a82e-47a304251ab5" containerName="barbican-api" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.476406 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="78cfb04c-4099-4211-a82e-47a304251ab5" containerName="barbican-api" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.476562 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cf4a1cf-ef96-4e50-946c-cd0c5eccff32" containerName="cinder-db-sync" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.476579 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="efca4ca3-8021-4417-83cc-1b26f13f97bd" containerName="dnsmasq-dns" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.476594 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="78cfb04c-4099-4211-a82e-47a304251ab5" containerName="barbican-api-log" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.476605 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="78cfb04c-4099-4211-a82e-47a304251ab5" containerName="barbican-api" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.484228 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bwn7\" (UniqueName: \"kubernetes.io/projected/78cfb04c-4099-4211-a82e-47a304251ab5-kube-api-access-6bwn7\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.484253 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78cfb04c-4099-4211-a82e-47a304251ab5-logs\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.484274 4813 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78cfb04c-4099-4211-a82e-47a304251ab5-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.499433 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.518248 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-wx499" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.518437 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.518572 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.518674 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.557390 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78cfb04c-4099-4211-a82e-47a304251ab5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "78cfb04c-4099-4211-a82e-47a304251ab5" (UID: "78cfb04c-4099-4211-a82e-47a304251ab5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.564392 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78cfb04c-4099-4211-a82e-47a304251ab5-config-data" (OuterVolumeSpecName: "config-data") pod "78cfb04c-4099-4211-a82e-47a304251ab5" (UID: "78cfb04c-4099-4211-a82e-47a304251ab5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.565186 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.585148 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77\") " pod="openstack/cinder-scheduler-0" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.585189 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77\") " pod="openstack/cinder-scheduler-0" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.585206 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-scripts\") pod \"cinder-scheduler-0\" (UID: \"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77\") " pod="openstack/cinder-scheduler-0" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.585276 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-config-data\") pod \"cinder-scheduler-0\" (UID: \"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77\") " pod="openstack/cinder-scheduler-0" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.585307 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccfdp\" (UniqueName: \"kubernetes.io/projected/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-kube-api-access-ccfdp\") pod \"cinder-scheduler-0\" (UID: \"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77\") " pod="openstack/cinder-scheduler-0" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.585369 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77\") " pod="openstack/cinder-scheduler-0" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.585473 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78cfb04c-4099-4211-a82e-47a304251ab5-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.585499 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78cfb04c-4099-4211-a82e-47a304251ab5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.634016 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-j49d2"] Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.635724 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-j49d2" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.671761 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-j49d2"] Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.679719 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" event={"ID":"efca4ca3-8021-4417-83cc-1b26f13f97bd","Type":"ContainerDied","Data":"d4857fdf061b281100b9d4b234170eab196bc39a7280eb9072f484773c83bc11"} Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.679785 4813 scope.go:117] "RemoveContainer" containerID="2b0d09d239154fc3944e334774798b0da04c92d93b83aee3bbb341b976d5155c" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.679938 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-rrfrw" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.693214 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f539e8ab-300c-426b-98e7-f9e87a6787b5" containerName="ceilometer-notification-agent" containerID="cri-o://dfbc16c809c9610b60841cb1e75b5a4aa13ade73ea1b1707f16b77a327c14418" gracePeriod=30 Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.693462 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f539e8ab-300c-426b-98e7-f9e87a6787b5","Type":"ContainerStarted","Data":"b1de04aae5c9f022eba72be9e7f4ce2b1e9f55e8628ba9553044885acbf66ea2"} Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.693510 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.693560 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f539e8ab-300c-426b-98e7-f9e87a6787b5" containerName="proxy-httpd" containerID="cri-o://b1de04aae5c9f022eba72be9e7f4ce2b1e9f55e8628ba9553044885acbf66ea2" gracePeriod=30 Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.693604 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f539e8ab-300c-426b-98e7-f9e87a6787b5" containerName="sg-core" containerID="cri-o://cf02350d626d2d9a7dd8311ef157f7dd95a5cbcbf89c6fccc653e5627ca1f6cd" gracePeriod=30 Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.693946 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77\") " pod="openstack/cinder-scheduler-0" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.694015 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77\") " pod="openstack/cinder-scheduler-0" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.694041 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77\") " pod="openstack/cinder-scheduler-0" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.694057 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-scripts\") pod \"cinder-scheduler-0\" (UID: \"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77\") " pod="openstack/cinder-scheduler-0" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.694138 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-config-data\") pod \"cinder-scheduler-0\" (UID: \"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77\") " pod="openstack/cinder-scheduler-0" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.694191 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccfdp\" (UniqueName: \"kubernetes.io/projected/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-kube-api-access-ccfdp\") pod \"cinder-scheduler-0\" (UID: \"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77\") " pod="openstack/cinder-scheduler-0" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.696350 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77\") " pod="openstack/cinder-scheduler-0" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.711497 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77\") " pod="openstack/cinder-scheduler-0" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.748416 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77\") " pod="openstack/cinder-scheduler-0" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.748588 4813 scope.go:117] "RemoveContainer" containerID="15ed9e2cd1cd900278014cb261390fa644c9ac32bac10da79943a454e74702cd" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.748788 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-scripts\") pod \"cinder-scheduler-0\" (UID: \"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77\") " pod="openstack/cinder-scheduler-0" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.749091 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-config-data\") pod \"cinder-scheduler-0\" (UID: \"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77\") " pod="openstack/cinder-scheduler-0" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.757858 4813 generic.go:334] "Generic (PLEG): container finished" podID="78cfb04c-4099-4211-a82e-47a304251ab5" containerID="acc4f465e1678852c5cec3f2ce1dda35b19bb8e717db2826e371bd238aef5f09" exitCode=0 Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.757906 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-66f946f97d-gh5xc" event={"ID":"78cfb04c-4099-4211-a82e-47a304251ab5","Type":"ContainerDied","Data":"acc4f465e1678852c5cec3f2ce1dda35b19bb8e717db2826e371bd238aef5f09"} Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.757933 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-66f946f97d-gh5xc" event={"ID":"78cfb04c-4099-4211-a82e-47a304251ab5","Type":"ContainerDied","Data":"1a239a6a5b6ee4d213a60af0e3d2c40cd7a6db930283f98e9065fa779c7d3a86"} Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.758001 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-66f946f97d-gh5xc" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.808080 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7e7e469-ed0d-463c-839c-95427c632d5b-config\") pod \"dnsmasq-dns-6578955fd5-j49d2\" (UID: \"d7e7e469-ed0d-463c-839c-95427c632d5b\") " pod="openstack/dnsmasq-dns-6578955fd5-j49d2" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.808200 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7e7e469-ed0d-463c-839c-95427c632d5b-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-j49d2\" (UID: \"d7e7e469-ed0d-463c-839c-95427c632d5b\") " pod="openstack/dnsmasq-dns-6578955fd5-j49d2" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.808333 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7e7e469-ed0d-463c-839c-95427c632d5b-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-j49d2\" (UID: \"d7e7e469-ed0d-463c-839c-95427c632d5b\") " pod="openstack/dnsmasq-dns-6578955fd5-j49d2" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.808365 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7e7e469-ed0d-463c-839c-95427c632d5b-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-j49d2\" (UID: \"d7e7e469-ed0d-463c-839c-95427c632d5b\") " pod="openstack/dnsmasq-dns-6578955fd5-j49d2" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.808489 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rj9bs\" (UniqueName: \"kubernetes.io/projected/d7e7e469-ed0d-463c-839c-95427c632d5b-kube-api-access-rj9bs\") pod \"dnsmasq-dns-6578955fd5-j49d2\" (UID: \"d7e7e469-ed0d-463c-839c-95427c632d5b\") " pod="openstack/dnsmasq-dns-6578955fd5-j49d2" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.808552 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7e7e469-ed0d-463c-839c-95427c632d5b-dns-svc\") pod \"dnsmasq-dns-6578955fd5-j49d2\" (UID: \"d7e7e469-ed0d-463c-839c-95427c632d5b\") " pod="openstack/dnsmasq-dns-6578955fd5-j49d2" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.813825 4813 scope.go:117] "RemoveContainer" containerID="acc4f465e1678852c5cec3f2ce1dda35b19bb8e717db2826e371bd238aef5f09" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.850998 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccfdp\" (UniqueName: \"kubernetes.io/projected/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-kube-api-access-ccfdp\") pod \"cinder-scheduler-0\" (UID: \"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77\") " pod="openstack/cinder-scheduler-0" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.851077 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-rrfrw"] Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.887981 4813 scope.go:117] "RemoveContainer" containerID="28a8a23ceae65c8b9be7f360a346a712fe0b76eaa4ec1db1bad715488651fd35" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.888011 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.910232 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7e7e469-ed0d-463c-839c-95427c632d5b-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-j49d2\" (UID: \"d7e7e469-ed0d-463c-839c-95427c632d5b\") " pod="openstack/dnsmasq-dns-6578955fd5-j49d2" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.910338 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7e7e469-ed0d-463c-839c-95427c632d5b-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-j49d2\" (UID: \"d7e7e469-ed0d-463c-839c-95427c632d5b\") " pod="openstack/dnsmasq-dns-6578955fd5-j49d2" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.910375 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7e7e469-ed0d-463c-839c-95427c632d5b-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-j49d2\" (UID: \"d7e7e469-ed0d-463c-839c-95427c632d5b\") " pod="openstack/dnsmasq-dns-6578955fd5-j49d2" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.910437 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rj9bs\" (UniqueName: \"kubernetes.io/projected/d7e7e469-ed0d-463c-839c-95427c632d5b-kube-api-access-rj9bs\") pod \"dnsmasq-dns-6578955fd5-j49d2\" (UID: \"d7e7e469-ed0d-463c-839c-95427c632d5b\") " pod="openstack/dnsmasq-dns-6578955fd5-j49d2" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.910468 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7e7e469-ed0d-463c-839c-95427c632d5b-dns-svc\") pod \"dnsmasq-dns-6578955fd5-j49d2\" (UID: \"d7e7e469-ed0d-463c-839c-95427c632d5b\") " pod="openstack/dnsmasq-dns-6578955fd5-j49d2" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.910494 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7e7e469-ed0d-463c-839c-95427c632d5b-config\") pod \"dnsmasq-dns-6578955fd5-j49d2\" (UID: \"d7e7e469-ed0d-463c-839c-95427c632d5b\") " pod="openstack/dnsmasq-dns-6578955fd5-j49d2" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.911484 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7e7e469-ed0d-463c-839c-95427c632d5b-config\") pod \"dnsmasq-dns-6578955fd5-j49d2\" (UID: \"d7e7e469-ed0d-463c-839c-95427c632d5b\") " pod="openstack/dnsmasq-dns-6578955fd5-j49d2" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.911997 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7e7e469-ed0d-463c-839c-95427c632d5b-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-j49d2\" (UID: \"d7e7e469-ed0d-463c-839c-95427c632d5b\") " pod="openstack/dnsmasq-dns-6578955fd5-j49d2" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.912494 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7e7e469-ed0d-463c-839c-95427c632d5b-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-j49d2\" (UID: \"d7e7e469-ed0d-463c-839c-95427c632d5b\") " pod="openstack/dnsmasq-dns-6578955fd5-j49d2" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.912853 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-rrfrw"] Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.930328 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-66f946f97d-gh5xc"] Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.930823 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7e7e469-ed0d-463c-839c-95427c632d5b-dns-svc\") pod \"dnsmasq-dns-6578955fd5-j49d2\" (UID: \"d7e7e469-ed0d-463c-839c-95427c632d5b\") " pod="openstack/dnsmasq-dns-6578955fd5-j49d2" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.930829 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7e7e469-ed0d-463c-839c-95427c632d5b-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-j49d2\" (UID: \"d7e7e469-ed0d-463c-839c-95427c632d5b\") " pod="openstack/dnsmasq-dns-6578955fd5-j49d2" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.938704 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rj9bs\" (UniqueName: \"kubernetes.io/projected/d7e7e469-ed0d-463c-839c-95427c632d5b-kube-api-access-rj9bs\") pod \"dnsmasq-dns-6578955fd5-j49d2\" (UID: \"d7e7e469-ed0d-463c-839c-95427c632d5b\") " pod="openstack/dnsmasq-dns-6578955fd5-j49d2" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.953373 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-66f946f97d-gh5xc"] Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.969715 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-j49d2" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.970142 4813 scope.go:117] "RemoveContainer" containerID="acc4f465e1678852c5cec3f2ce1dda35b19bb8e717db2826e371bd238aef5f09" Dec 06 16:03:48 crc kubenswrapper[4813]: E1206 16:03:48.975365 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acc4f465e1678852c5cec3f2ce1dda35b19bb8e717db2826e371bd238aef5f09\": container with ID starting with acc4f465e1678852c5cec3f2ce1dda35b19bb8e717db2826e371bd238aef5f09 not found: ID does not exist" containerID="acc4f465e1678852c5cec3f2ce1dda35b19bb8e717db2826e371bd238aef5f09" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.975398 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acc4f465e1678852c5cec3f2ce1dda35b19bb8e717db2826e371bd238aef5f09"} err="failed to get container status \"acc4f465e1678852c5cec3f2ce1dda35b19bb8e717db2826e371bd238aef5f09\": rpc error: code = NotFound desc = could not find container \"acc4f465e1678852c5cec3f2ce1dda35b19bb8e717db2826e371bd238aef5f09\": container with ID starting with acc4f465e1678852c5cec3f2ce1dda35b19bb8e717db2826e371bd238aef5f09 not found: ID does not exist" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.975420 4813 scope.go:117] "RemoveContainer" containerID="28a8a23ceae65c8b9be7f360a346a712fe0b76eaa4ec1db1bad715488651fd35" Dec 06 16:03:48 crc kubenswrapper[4813]: E1206 16:03:48.975648 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28a8a23ceae65c8b9be7f360a346a712fe0b76eaa4ec1db1bad715488651fd35\": container with ID starting with 28a8a23ceae65c8b9be7f360a346a712fe0b76eaa4ec1db1bad715488651fd35 not found: ID does not exist" containerID="28a8a23ceae65c8b9be7f360a346a712fe0b76eaa4ec1db1bad715488651fd35" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.975667 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28a8a23ceae65c8b9be7f360a346a712fe0b76eaa4ec1db1bad715488651fd35"} err="failed to get container status \"28a8a23ceae65c8b9be7f360a346a712fe0b76eaa4ec1db1bad715488651fd35\": rpc error: code = NotFound desc = could not find container \"28a8a23ceae65c8b9be7f360a346a712fe0b76eaa4ec1db1bad715488651fd35\": container with ID starting with 28a8a23ceae65c8b9be7f360a346a712fe0b76eaa4ec1db1bad715488651fd35 not found: ID does not exist" Dec 06 16:03:48 crc kubenswrapper[4813]: I1206 16:03:48.993573 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-759b8b45fc-psf25" Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.033659 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.035111 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.044656 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.091152 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.117537 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7c8fc98b5b-dwkdc"] Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.117736 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7c8fc98b5b-dwkdc" podUID="7c583806-afc7-4f95-8d4c-f0690157ca65" containerName="neutron-api" containerID="cri-o://9999291e2b8d7565b054e93fa369aa0a9392136209dae8df114b56c191ed251a" gracePeriod=30 Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.117865 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7c8fc98b5b-dwkdc" podUID="7c583806-afc7-4f95-8d4c-f0690157ca65" containerName="neutron-httpd" containerID="cri-o://d6326e7e0b145e3ecd7b906844f31d37d84c0c433cced12074a74eecb1eb6d0d" gracePeriod=30 Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.221571 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78n6q\" (UniqueName: \"kubernetes.io/projected/655286a9-c261-43f2-ae31-80222cd3883b-kube-api-access-78n6q\") pod \"cinder-api-0\" (UID: \"655286a9-c261-43f2-ae31-80222cd3883b\") " pod="openstack/cinder-api-0" Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.221630 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/655286a9-c261-43f2-ae31-80222cd3883b-config-data\") pod \"cinder-api-0\" (UID: \"655286a9-c261-43f2-ae31-80222cd3883b\") " pod="openstack/cinder-api-0" Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.221661 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/655286a9-c261-43f2-ae31-80222cd3883b-config-data-custom\") pod \"cinder-api-0\" (UID: \"655286a9-c261-43f2-ae31-80222cd3883b\") " pod="openstack/cinder-api-0" Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.221707 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/655286a9-c261-43f2-ae31-80222cd3883b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"655286a9-c261-43f2-ae31-80222cd3883b\") " pod="openstack/cinder-api-0" Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.221725 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/655286a9-c261-43f2-ae31-80222cd3883b-scripts\") pod \"cinder-api-0\" (UID: \"655286a9-c261-43f2-ae31-80222cd3883b\") " pod="openstack/cinder-api-0" Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.221757 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/655286a9-c261-43f2-ae31-80222cd3883b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"655286a9-c261-43f2-ae31-80222cd3883b\") " pod="openstack/cinder-api-0" Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.221789 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/655286a9-c261-43f2-ae31-80222cd3883b-logs\") pod \"cinder-api-0\" (UID: \"655286a9-c261-43f2-ae31-80222cd3883b\") " pod="openstack/cinder-api-0" Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.327985 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/655286a9-c261-43f2-ae31-80222cd3883b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"655286a9-c261-43f2-ae31-80222cd3883b\") " pod="openstack/cinder-api-0" Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.328288 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/655286a9-c261-43f2-ae31-80222cd3883b-scripts\") pod \"cinder-api-0\" (UID: \"655286a9-c261-43f2-ae31-80222cd3883b\") " pod="openstack/cinder-api-0" Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.328323 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/655286a9-c261-43f2-ae31-80222cd3883b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"655286a9-c261-43f2-ae31-80222cd3883b\") " pod="openstack/cinder-api-0" Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.328374 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/655286a9-c261-43f2-ae31-80222cd3883b-logs\") pod \"cinder-api-0\" (UID: \"655286a9-c261-43f2-ae31-80222cd3883b\") " pod="openstack/cinder-api-0" Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.328416 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78n6q\" (UniqueName: \"kubernetes.io/projected/655286a9-c261-43f2-ae31-80222cd3883b-kube-api-access-78n6q\") pod \"cinder-api-0\" (UID: \"655286a9-c261-43f2-ae31-80222cd3883b\") " pod="openstack/cinder-api-0" Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.328447 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/655286a9-c261-43f2-ae31-80222cd3883b-config-data\") pod \"cinder-api-0\" (UID: \"655286a9-c261-43f2-ae31-80222cd3883b\") " pod="openstack/cinder-api-0" Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.328479 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/655286a9-c261-43f2-ae31-80222cd3883b-config-data-custom\") pod \"cinder-api-0\" (UID: \"655286a9-c261-43f2-ae31-80222cd3883b\") " pod="openstack/cinder-api-0" Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.334200 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/655286a9-c261-43f2-ae31-80222cd3883b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"655286a9-c261-43f2-ae31-80222cd3883b\") " pod="openstack/cinder-api-0" Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.334271 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/655286a9-c261-43f2-ae31-80222cd3883b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"655286a9-c261-43f2-ae31-80222cd3883b\") " pod="openstack/cinder-api-0" Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.334502 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/655286a9-c261-43f2-ae31-80222cd3883b-logs\") pod \"cinder-api-0\" (UID: \"655286a9-c261-43f2-ae31-80222cd3883b\") " pod="openstack/cinder-api-0" Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.336664 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/655286a9-c261-43f2-ae31-80222cd3883b-scripts\") pod \"cinder-api-0\" (UID: \"655286a9-c261-43f2-ae31-80222cd3883b\") " pod="openstack/cinder-api-0" Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.347723 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/655286a9-c261-43f2-ae31-80222cd3883b-config-data-custom\") pod \"cinder-api-0\" (UID: \"655286a9-c261-43f2-ae31-80222cd3883b\") " pod="openstack/cinder-api-0" Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.348137 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/655286a9-c261-43f2-ae31-80222cd3883b-config-data\") pod \"cinder-api-0\" (UID: \"655286a9-c261-43f2-ae31-80222cd3883b\") " pod="openstack/cinder-api-0" Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.370919 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78n6q\" (UniqueName: \"kubernetes.io/projected/655286a9-c261-43f2-ae31-80222cd3883b-kube-api-access-78n6q\") pod \"cinder-api-0\" (UID: \"655286a9-c261-43f2-ae31-80222cd3883b\") " pod="openstack/cinder-api-0" Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.398667 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.500394 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.769046 4813 generic.go:334] "Generic (PLEG): container finished" podID="f539e8ab-300c-426b-98e7-f9e87a6787b5" containerID="b1de04aae5c9f022eba72be9e7f4ce2b1e9f55e8628ba9553044885acbf66ea2" exitCode=0 Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.769370 4813 generic.go:334] "Generic (PLEG): container finished" podID="f539e8ab-300c-426b-98e7-f9e87a6787b5" containerID="cf02350d626d2d9a7dd8311ef157f7dd95a5cbcbf89c6fccc653e5627ca1f6cd" exitCode=2 Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.769405 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f539e8ab-300c-426b-98e7-f9e87a6787b5","Type":"ContainerDied","Data":"b1de04aae5c9f022eba72be9e7f4ce2b1e9f55e8628ba9553044885acbf66ea2"} Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.769426 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f539e8ab-300c-426b-98e7-f9e87a6787b5","Type":"ContainerDied","Data":"cf02350d626d2d9a7dd8311ef157f7dd95a5cbcbf89c6fccc653e5627ca1f6cd"} Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.771402 4813 generic.go:334] "Generic (PLEG): container finished" podID="7c583806-afc7-4f95-8d4c-f0690157ca65" containerID="d6326e7e0b145e3ecd7b906844f31d37d84c0c433cced12074a74eecb1eb6d0d" exitCode=0 Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.771448 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7c8fc98b5b-dwkdc" event={"ID":"7c583806-afc7-4f95-8d4c-f0690157ca65","Type":"ContainerDied","Data":"d6326e7e0b145e3ecd7b906844f31d37d84c0c433cced12074a74eecb1eb6d0d"} Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.772719 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77","Type":"ContainerStarted","Data":"63fa3e880fd7361099030f8e2114673965cfc912c1bedbf25f3435a16306fc7d"} Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.861488 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-j49d2"] Dec 06 16:03:49 crc kubenswrapper[4813]: W1206 16:03:49.864021 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7e7e469_ed0d_463c_839c_95427c632d5b.slice/crio-a58a45b0f7b1e35e62105d2e77c5e9fb0706feab2dddd9ec0990f6066eb0c8e9 WatchSource:0}: Error finding container a58a45b0f7b1e35e62105d2e77c5e9fb0706feab2dddd9ec0990f6066eb0c8e9: Status 404 returned error can't find the container with id a58a45b0f7b1e35e62105d2e77c5e9fb0706feab2dddd9ec0990f6066eb0c8e9 Dec 06 16:03:49 crc kubenswrapper[4813]: I1206 16:03:49.984295 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 06 16:03:50 crc kubenswrapper[4813]: I1206 16:03:50.504360 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78cfb04c-4099-4211-a82e-47a304251ab5" path="/var/lib/kubelet/pods/78cfb04c-4099-4211-a82e-47a304251ab5/volumes" Dec 06 16:03:50 crc kubenswrapper[4813]: I1206 16:03:50.505216 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efca4ca3-8021-4417-83cc-1b26f13f97bd" path="/var/lib/kubelet/pods/efca4ca3-8021-4417-83cc-1b26f13f97bd/volumes" Dec 06 16:03:50 crc kubenswrapper[4813]: I1206 16:03:50.794483 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"655286a9-c261-43f2-ae31-80222cd3883b","Type":"ContainerStarted","Data":"dfcc1517170cbb74cb8514cb7d7a7b48eec334b8fb1a10133a93f4989c001b8e"} Dec 06 16:03:50 crc kubenswrapper[4813]: I1206 16:03:50.805389 4813 generic.go:334] "Generic (PLEG): container finished" podID="d7e7e469-ed0d-463c-839c-95427c632d5b" containerID="ced8f07e40e0ca4da9ceb3d788b88b7be92cd599726686b994d7eb1f2895313a" exitCode=0 Dec 06 16:03:50 crc kubenswrapper[4813]: I1206 16:03:50.805437 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-j49d2" event={"ID":"d7e7e469-ed0d-463c-839c-95427c632d5b","Type":"ContainerDied","Data":"ced8f07e40e0ca4da9ceb3d788b88b7be92cd599726686b994d7eb1f2895313a"} Dec 06 16:03:50 crc kubenswrapper[4813]: I1206 16:03:50.805461 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-j49d2" event={"ID":"d7e7e469-ed0d-463c-839c-95427c632d5b","Type":"ContainerStarted","Data":"a58a45b0f7b1e35e62105d2e77c5e9fb0706feab2dddd9ec0990f6066eb0c8e9"} Dec 06 16:03:51 crc kubenswrapper[4813]: I1206 16:03:51.382112 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 06 16:03:51 crc kubenswrapper[4813]: I1206 16:03:51.835876 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-j49d2" event={"ID":"d7e7e469-ed0d-463c-839c-95427c632d5b","Type":"ContainerStarted","Data":"f1184bf09f3b6bba16396cf9b3e4d07ce090b43676ce244963ac2cd8f8eddd31"} Dec 06 16:03:51 crc kubenswrapper[4813]: I1206 16:03:51.836165 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6578955fd5-j49d2" Dec 06 16:03:51 crc kubenswrapper[4813]: I1206 16:03:51.838191 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77","Type":"ContainerStarted","Data":"b3dad93bf75be6fdd022ed5d1d3a707a617523b4c6cbd751f73b7790327f0e38"} Dec 06 16:03:51 crc kubenswrapper[4813]: I1206 16:03:51.847484 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"655286a9-c261-43f2-ae31-80222cd3883b","Type":"ContainerStarted","Data":"b239f9f8021d81b02a31ed98ebe03f0e963b164ab802e62d7533cb120ff53001"} Dec 06 16:03:51 crc kubenswrapper[4813]: I1206 16:03:51.856869 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6578955fd5-j49d2" podStartSLOduration=3.856853567 podStartE2EDuration="3.856853567s" podCreationTimestamp="2025-12-06 16:03:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:03:51.852116713 +0000 UTC m=+1071.742996289" watchObservedRunningTime="2025-12-06 16:03:51.856853567 +0000 UTC m=+1071.747733143" Dec 06 16:03:51 crc kubenswrapper[4813]: I1206 16:03:51.860028 4813 generic.go:334] "Generic (PLEG): container finished" podID="f539e8ab-300c-426b-98e7-f9e87a6787b5" containerID="dfbc16c809c9610b60841cb1e75b5a4aa13ade73ea1b1707f16b77a327c14418" exitCode=0 Dec 06 16:03:51 crc kubenswrapper[4813]: I1206 16:03:51.860090 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f539e8ab-300c-426b-98e7-f9e87a6787b5","Type":"ContainerDied","Data":"dfbc16c809c9610b60841cb1e75b5a4aa13ade73ea1b1707f16b77a327c14418"} Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.166653 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.313325 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f539e8ab-300c-426b-98e7-f9e87a6787b5-sg-core-conf-yaml\") pod \"f539e8ab-300c-426b-98e7-f9e87a6787b5\" (UID: \"f539e8ab-300c-426b-98e7-f9e87a6787b5\") " Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.313397 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f539e8ab-300c-426b-98e7-f9e87a6787b5-run-httpd\") pod \"f539e8ab-300c-426b-98e7-f9e87a6787b5\" (UID: \"f539e8ab-300c-426b-98e7-f9e87a6787b5\") " Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.313415 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f539e8ab-300c-426b-98e7-f9e87a6787b5-scripts\") pod \"f539e8ab-300c-426b-98e7-f9e87a6787b5\" (UID: \"f539e8ab-300c-426b-98e7-f9e87a6787b5\") " Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.313460 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f539e8ab-300c-426b-98e7-f9e87a6787b5-combined-ca-bundle\") pod \"f539e8ab-300c-426b-98e7-f9e87a6787b5\" (UID: \"f539e8ab-300c-426b-98e7-f9e87a6787b5\") " Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.313528 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6tkg\" (UniqueName: \"kubernetes.io/projected/f539e8ab-300c-426b-98e7-f9e87a6787b5-kube-api-access-k6tkg\") pod \"f539e8ab-300c-426b-98e7-f9e87a6787b5\" (UID: \"f539e8ab-300c-426b-98e7-f9e87a6787b5\") " Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.313549 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f539e8ab-300c-426b-98e7-f9e87a6787b5-config-data\") pod \"f539e8ab-300c-426b-98e7-f9e87a6787b5\" (UID: \"f539e8ab-300c-426b-98e7-f9e87a6787b5\") " Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.313654 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f539e8ab-300c-426b-98e7-f9e87a6787b5-log-httpd\") pod \"f539e8ab-300c-426b-98e7-f9e87a6787b5\" (UID: \"f539e8ab-300c-426b-98e7-f9e87a6787b5\") " Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.313834 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f539e8ab-300c-426b-98e7-f9e87a6787b5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f539e8ab-300c-426b-98e7-f9e87a6787b5" (UID: "f539e8ab-300c-426b-98e7-f9e87a6787b5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.314030 4813 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f539e8ab-300c-426b-98e7-f9e87a6787b5-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.314338 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f539e8ab-300c-426b-98e7-f9e87a6787b5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f539e8ab-300c-426b-98e7-f9e87a6787b5" (UID: "f539e8ab-300c-426b-98e7-f9e87a6787b5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.330389 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f539e8ab-300c-426b-98e7-f9e87a6787b5-kube-api-access-k6tkg" (OuterVolumeSpecName: "kube-api-access-k6tkg") pod "f539e8ab-300c-426b-98e7-f9e87a6787b5" (UID: "f539e8ab-300c-426b-98e7-f9e87a6787b5"). InnerVolumeSpecName "kube-api-access-k6tkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.333383 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f539e8ab-300c-426b-98e7-f9e87a6787b5-scripts" (OuterVolumeSpecName: "scripts") pod "f539e8ab-300c-426b-98e7-f9e87a6787b5" (UID: "f539e8ab-300c-426b-98e7-f9e87a6787b5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.416395 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f539e8ab-300c-426b-98e7-f9e87a6787b5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f539e8ab-300c-426b-98e7-f9e87a6787b5" (UID: "f539e8ab-300c-426b-98e7-f9e87a6787b5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.416769 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f539e8ab-300c-426b-98e7-f9e87a6787b5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.416801 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6tkg\" (UniqueName: \"kubernetes.io/projected/f539e8ab-300c-426b-98e7-f9e87a6787b5-kube-api-access-k6tkg\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.416810 4813 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f539e8ab-300c-426b-98e7-f9e87a6787b5-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.416818 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f539e8ab-300c-426b-98e7-f9e87a6787b5-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.424399 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f539e8ab-300c-426b-98e7-f9e87a6787b5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f539e8ab-300c-426b-98e7-f9e87a6787b5" (UID: "f539e8ab-300c-426b-98e7-f9e87a6787b5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.468420 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f539e8ab-300c-426b-98e7-f9e87a6787b5-config-data" (OuterVolumeSpecName: "config-data") pod "f539e8ab-300c-426b-98e7-f9e87a6787b5" (UID: "f539e8ab-300c-426b-98e7-f9e87a6787b5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.519802 4813 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f539e8ab-300c-426b-98e7-f9e87a6787b5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.520036 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f539e8ab-300c-426b-98e7-f9e87a6787b5-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.870293 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77","Type":"ContainerStarted","Data":"0eb7370c8ca2d86993b9b89e3aa1ac9a358a84d936bcf30ef9d764e36ca89976"} Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.872121 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="655286a9-c261-43f2-ae31-80222cd3883b" containerName="cinder-api-log" containerID="cri-o://b239f9f8021d81b02a31ed98ebe03f0e963b164ab802e62d7533cb120ff53001" gracePeriod=30 Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.872217 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"655286a9-c261-43f2-ae31-80222cd3883b","Type":"ContainerStarted","Data":"4aaa5a29ce607255506e9698e7e8f1147a16a75052db71bf069781dfe6e63cce"} Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.872158 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="655286a9-c261-43f2-ae31-80222cd3883b" containerName="cinder-api" containerID="cri-o://4aaa5a29ce607255506e9698e7e8f1147a16a75052db71bf069781dfe6e63cce" gracePeriod=30 Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.872300 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.877244 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f539e8ab-300c-426b-98e7-f9e87a6787b5","Type":"ContainerDied","Data":"970e7c7a4af6f21c2152801c05d4c15dfb412dea047584ad9886f3dde225be9c"} Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.877312 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.877318 4813 scope.go:117] "RemoveContainer" containerID="b1de04aae5c9f022eba72be9e7f4ce2b1e9f55e8628ba9553044885acbf66ea2" Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.893958 4813 scope.go:117] "RemoveContainer" containerID="cf02350d626d2d9a7dd8311ef157f7dd95a5cbcbf89c6fccc653e5627ca1f6cd" Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.920752 4813 scope.go:117] "RemoveContainer" containerID="dfbc16c809c9610b60841cb1e75b5a4aa13ade73ea1b1707f16b77a327c14418" Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.933700 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.281228095 podStartE2EDuration="4.933684715s" podCreationTimestamp="2025-12-06 16:03:48 +0000 UTC" firstStartedPulling="2025-12-06 16:03:49.517695449 +0000 UTC m=+1069.408575015" lastFinishedPulling="2025-12-06 16:03:50.170152059 +0000 UTC m=+1070.061031635" observedRunningTime="2025-12-06 16:03:52.903999695 +0000 UTC m=+1072.794879281" watchObservedRunningTime="2025-12-06 16:03:52.933684715 +0000 UTC m=+1072.824564291" Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.947445 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.955576 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.980987 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:03:52 crc kubenswrapper[4813]: E1206 16:03:52.981366 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f539e8ab-300c-426b-98e7-f9e87a6787b5" containerName="ceilometer-notification-agent" Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.981386 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f539e8ab-300c-426b-98e7-f9e87a6787b5" containerName="ceilometer-notification-agent" Dec 06 16:03:52 crc kubenswrapper[4813]: E1206 16:03:52.981422 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f539e8ab-300c-426b-98e7-f9e87a6787b5" containerName="proxy-httpd" Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.981431 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f539e8ab-300c-426b-98e7-f9e87a6787b5" containerName="proxy-httpd" Dec 06 16:03:52 crc kubenswrapper[4813]: E1206 16:03:52.981453 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f539e8ab-300c-426b-98e7-f9e87a6787b5" containerName="sg-core" Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.981462 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f539e8ab-300c-426b-98e7-f9e87a6787b5" containerName="sg-core" Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.981649 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f539e8ab-300c-426b-98e7-f9e87a6787b5" containerName="sg-core" Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.981673 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f539e8ab-300c-426b-98e7-f9e87a6787b5" containerName="proxy-httpd" Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.981686 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f539e8ab-300c-426b-98e7-f9e87a6787b5" containerName="ceilometer-notification-agent" Dec 06 16:03:52 crc kubenswrapper[4813]: I1206 16:03:52.984471 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.000503 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.00048056 podStartE2EDuration="5.00048056s" podCreationTimestamp="2025-12-06 16:03:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:03:52.992446669 +0000 UTC m=+1072.883326245" watchObservedRunningTime="2025-12-06 16:03:53.00048056 +0000 UTC m=+1072.891360136" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.002813 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.003021 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.026917 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.140459 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f56eb906-c881-4529-8c56-f11d6246a7f5-run-httpd\") pod \"ceilometer-0\" (UID: \"f56eb906-c881-4529-8c56-f11d6246a7f5\") " pod="openstack/ceilometer-0" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.140520 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f56eb906-c881-4529-8c56-f11d6246a7f5-config-data\") pod \"ceilometer-0\" (UID: \"f56eb906-c881-4529-8c56-f11d6246a7f5\") " pod="openstack/ceilometer-0" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.140573 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kkfc\" (UniqueName: \"kubernetes.io/projected/f56eb906-c881-4529-8c56-f11d6246a7f5-kube-api-access-7kkfc\") pod \"ceilometer-0\" (UID: \"f56eb906-c881-4529-8c56-f11d6246a7f5\") " pod="openstack/ceilometer-0" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.140593 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f56eb906-c881-4529-8c56-f11d6246a7f5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f56eb906-c881-4529-8c56-f11d6246a7f5\") " pod="openstack/ceilometer-0" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.140647 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f56eb906-c881-4529-8c56-f11d6246a7f5-log-httpd\") pod \"ceilometer-0\" (UID: \"f56eb906-c881-4529-8c56-f11d6246a7f5\") " pod="openstack/ceilometer-0" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.140669 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f56eb906-c881-4529-8c56-f11d6246a7f5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f56eb906-c881-4529-8c56-f11d6246a7f5\") " pod="openstack/ceilometer-0" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.140693 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f56eb906-c881-4529-8c56-f11d6246a7f5-scripts\") pod \"ceilometer-0\" (UID: \"f56eb906-c881-4529-8c56-f11d6246a7f5\") " pod="openstack/ceilometer-0" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.241730 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f56eb906-c881-4529-8c56-f11d6246a7f5-run-httpd\") pod \"ceilometer-0\" (UID: \"f56eb906-c881-4529-8c56-f11d6246a7f5\") " pod="openstack/ceilometer-0" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.241789 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f56eb906-c881-4529-8c56-f11d6246a7f5-config-data\") pod \"ceilometer-0\" (UID: \"f56eb906-c881-4529-8c56-f11d6246a7f5\") " pod="openstack/ceilometer-0" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.241833 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kkfc\" (UniqueName: \"kubernetes.io/projected/f56eb906-c881-4529-8c56-f11d6246a7f5-kube-api-access-7kkfc\") pod \"ceilometer-0\" (UID: \"f56eb906-c881-4529-8c56-f11d6246a7f5\") " pod="openstack/ceilometer-0" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.241853 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f56eb906-c881-4529-8c56-f11d6246a7f5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f56eb906-c881-4529-8c56-f11d6246a7f5\") " pod="openstack/ceilometer-0" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.241895 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f56eb906-c881-4529-8c56-f11d6246a7f5-log-httpd\") pod \"ceilometer-0\" (UID: \"f56eb906-c881-4529-8c56-f11d6246a7f5\") " pod="openstack/ceilometer-0" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.241919 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f56eb906-c881-4529-8c56-f11d6246a7f5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f56eb906-c881-4529-8c56-f11d6246a7f5\") " pod="openstack/ceilometer-0" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.241941 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f56eb906-c881-4529-8c56-f11d6246a7f5-scripts\") pod \"ceilometer-0\" (UID: \"f56eb906-c881-4529-8c56-f11d6246a7f5\") " pod="openstack/ceilometer-0" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.242165 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f56eb906-c881-4529-8c56-f11d6246a7f5-run-httpd\") pod \"ceilometer-0\" (UID: \"f56eb906-c881-4529-8c56-f11d6246a7f5\") " pod="openstack/ceilometer-0" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.248653 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f56eb906-c881-4529-8c56-f11d6246a7f5-log-httpd\") pod \"ceilometer-0\" (UID: \"f56eb906-c881-4529-8c56-f11d6246a7f5\") " pod="openstack/ceilometer-0" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.253146 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f56eb906-c881-4529-8c56-f11d6246a7f5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f56eb906-c881-4529-8c56-f11d6246a7f5\") " pod="openstack/ceilometer-0" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.257461 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f56eb906-c881-4529-8c56-f11d6246a7f5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f56eb906-c881-4529-8c56-f11d6246a7f5\") " pod="openstack/ceilometer-0" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.258079 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f56eb906-c881-4529-8c56-f11d6246a7f5-scripts\") pod \"ceilometer-0\" (UID: \"f56eb906-c881-4529-8c56-f11d6246a7f5\") " pod="openstack/ceilometer-0" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.260821 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f56eb906-c881-4529-8c56-f11d6246a7f5-config-data\") pod \"ceilometer-0\" (UID: \"f56eb906-c881-4529-8c56-f11d6246a7f5\") " pod="openstack/ceilometer-0" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.274018 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kkfc\" (UniqueName: \"kubernetes.io/projected/f56eb906-c881-4529-8c56-f11d6246a7f5-kube-api-access-7kkfc\") pod \"ceilometer-0\" (UID: \"f56eb906-c881-4529-8c56-f11d6246a7f5\") " pod="openstack/ceilometer-0" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.378102 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.850662 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.888114 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.893515 4813 generic.go:334] "Generic (PLEG): container finished" podID="655286a9-c261-43f2-ae31-80222cd3883b" containerID="4aaa5a29ce607255506e9698e7e8f1147a16a75052db71bf069781dfe6e63cce" exitCode=0 Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.893544 4813 generic.go:334] "Generic (PLEG): container finished" podID="655286a9-c261-43f2-ae31-80222cd3883b" containerID="b239f9f8021d81b02a31ed98ebe03f0e963b164ab802e62d7533cb120ff53001" exitCode=143 Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.893582 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"655286a9-c261-43f2-ae31-80222cd3883b","Type":"ContainerDied","Data":"4aaa5a29ce607255506e9698e7e8f1147a16a75052db71bf069781dfe6e63cce"} Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.893609 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"655286a9-c261-43f2-ae31-80222cd3883b","Type":"ContainerDied","Data":"b239f9f8021d81b02a31ed98ebe03f0e963b164ab802e62d7533cb120ff53001"} Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.893619 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"655286a9-c261-43f2-ae31-80222cd3883b","Type":"ContainerDied","Data":"dfcc1517170cbb74cb8514cb7d7a7b48eec334b8fb1a10133a93f4989c001b8e"} Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.893635 4813 scope.go:117] "RemoveContainer" containerID="4aaa5a29ce607255506e9698e7e8f1147a16a75052db71bf069781dfe6e63cce" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.893719 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.917918 4813 scope.go:117] "RemoveContainer" containerID="b239f9f8021d81b02a31ed98ebe03f0e963b164ab802e62d7533cb120ff53001" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.954466 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/655286a9-c261-43f2-ae31-80222cd3883b-scripts\") pod \"655286a9-c261-43f2-ae31-80222cd3883b\" (UID: \"655286a9-c261-43f2-ae31-80222cd3883b\") " Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.954676 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78n6q\" (UniqueName: \"kubernetes.io/projected/655286a9-c261-43f2-ae31-80222cd3883b-kube-api-access-78n6q\") pod \"655286a9-c261-43f2-ae31-80222cd3883b\" (UID: \"655286a9-c261-43f2-ae31-80222cd3883b\") " Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.954768 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/655286a9-c261-43f2-ae31-80222cd3883b-logs\") pod \"655286a9-c261-43f2-ae31-80222cd3883b\" (UID: \"655286a9-c261-43f2-ae31-80222cd3883b\") " Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.954799 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/655286a9-c261-43f2-ae31-80222cd3883b-config-data\") pod \"655286a9-c261-43f2-ae31-80222cd3883b\" (UID: \"655286a9-c261-43f2-ae31-80222cd3883b\") " Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.954846 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/655286a9-c261-43f2-ae31-80222cd3883b-config-data-custom\") pod \"655286a9-c261-43f2-ae31-80222cd3883b\" (UID: \"655286a9-c261-43f2-ae31-80222cd3883b\") " Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.954862 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/655286a9-c261-43f2-ae31-80222cd3883b-etc-machine-id\") pod \"655286a9-c261-43f2-ae31-80222cd3883b\" (UID: \"655286a9-c261-43f2-ae31-80222cd3883b\") " Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.954916 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/655286a9-c261-43f2-ae31-80222cd3883b-combined-ca-bundle\") pod \"655286a9-c261-43f2-ae31-80222cd3883b\" (UID: \"655286a9-c261-43f2-ae31-80222cd3883b\") " Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.955135 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/655286a9-c261-43f2-ae31-80222cd3883b-logs" (OuterVolumeSpecName: "logs") pod "655286a9-c261-43f2-ae31-80222cd3883b" (UID: "655286a9-c261-43f2-ae31-80222cd3883b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.955493 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/655286a9-c261-43f2-ae31-80222cd3883b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "655286a9-c261-43f2-ae31-80222cd3883b" (UID: "655286a9-c261-43f2-ae31-80222cd3883b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.955613 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/655286a9-c261-43f2-ae31-80222cd3883b-logs\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.955628 4813 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/655286a9-c261-43f2-ae31-80222cd3883b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.955681 4813 scope.go:117] "RemoveContainer" containerID="4aaa5a29ce607255506e9698e7e8f1147a16a75052db71bf069781dfe6e63cce" Dec 06 16:03:53 crc kubenswrapper[4813]: E1206 16:03:53.956650 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4aaa5a29ce607255506e9698e7e8f1147a16a75052db71bf069781dfe6e63cce\": container with ID starting with 4aaa5a29ce607255506e9698e7e8f1147a16a75052db71bf069781dfe6e63cce not found: ID does not exist" containerID="4aaa5a29ce607255506e9698e7e8f1147a16a75052db71bf069781dfe6e63cce" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.956672 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4aaa5a29ce607255506e9698e7e8f1147a16a75052db71bf069781dfe6e63cce"} err="failed to get container status \"4aaa5a29ce607255506e9698e7e8f1147a16a75052db71bf069781dfe6e63cce\": rpc error: code = NotFound desc = could not find container \"4aaa5a29ce607255506e9698e7e8f1147a16a75052db71bf069781dfe6e63cce\": container with ID starting with 4aaa5a29ce607255506e9698e7e8f1147a16a75052db71bf069781dfe6e63cce not found: ID does not exist" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.956707 4813 scope.go:117] "RemoveContainer" containerID="b239f9f8021d81b02a31ed98ebe03f0e963b164ab802e62d7533cb120ff53001" Dec 06 16:03:53 crc kubenswrapper[4813]: E1206 16:03:53.957013 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b239f9f8021d81b02a31ed98ebe03f0e963b164ab802e62d7533cb120ff53001\": container with ID starting with b239f9f8021d81b02a31ed98ebe03f0e963b164ab802e62d7533cb120ff53001 not found: ID does not exist" containerID="b239f9f8021d81b02a31ed98ebe03f0e963b164ab802e62d7533cb120ff53001" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.957030 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b239f9f8021d81b02a31ed98ebe03f0e963b164ab802e62d7533cb120ff53001"} err="failed to get container status \"b239f9f8021d81b02a31ed98ebe03f0e963b164ab802e62d7533cb120ff53001\": rpc error: code = NotFound desc = could not find container \"b239f9f8021d81b02a31ed98ebe03f0e963b164ab802e62d7533cb120ff53001\": container with ID starting with b239f9f8021d81b02a31ed98ebe03f0e963b164ab802e62d7533cb120ff53001 not found: ID does not exist" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.957043 4813 scope.go:117] "RemoveContainer" containerID="4aaa5a29ce607255506e9698e7e8f1147a16a75052db71bf069781dfe6e63cce" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.957252 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4aaa5a29ce607255506e9698e7e8f1147a16a75052db71bf069781dfe6e63cce"} err="failed to get container status \"4aaa5a29ce607255506e9698e7e8f1147a16a75052db71bf069781dfe6e63cce\": rpc error: code = NotFound desc = could not find container \"4aaa5a29ce607255506e9698e7e8f1147a16a75052db71bf069781dfe6e63cce\": container with ID starting with 4aaa5a29ce607255506e9698e7e8f1147a16a75052db71bf069781dfe6e63cce not found: ID does not exist" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.957276 4813 scope.go:117] "RemoveContainer" containerID="b239f9f8021d81b02a31ed98ebe03f0e963b164ab802e62d7533cb120ff53001" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.957418 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b239f9f8021d81b02a31ed98ebe03f0e963b164ab802e62d7533cb120ff53001"} err="failed to get container status \"b239f9f8021d81b02a31ed98ebe03f0e963b164ab802e62d7533cb120ff53001\": rpc error: code = NotFound desc = could not find container \"b239f9f8021d81b02a31ed98ebe03f0e963b164ab802e62d7533cb120ff53001\": container with ID starting with b239f9f8021d81b02a31ed98ebe03f0e963b164ab802e62d7533cb120ff53001 not found: ID does not exist" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.963547 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/655286a9-c261-43f2-ae31-80222cd3883b-kube-api-access-78n6q" (OuterVolumeSpecName: "kube-api-access-78n6q") pod "655286a9-c261-43f2-ae31-80222cd3883b" (UID: "655286a9-c261-43f2-ae31-80222cd3883b"). InnerVolumeSpecName "kube-api-access-78n6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.970386 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/655286a9-c261-43f2-ae31-80222cd3883b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "655286a9-c261-43f2-ae31-80222cd3883b" (UID: "655286a9-c261-43f2-ae31-80222cd3883b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.974998 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:03:53 crc kubenswrapper[4813]: I1206 16:03:53.982195 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/655286a9-c261-43f2-ae31-80222cd3883b-scripts" (OuterVolumeSpecName: "scripts") pod "655286a9-c261-43f2-ae31-80222cd3883b" (UID: "655286a9-c261-43f2-ae31-80222cd3883b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.006394 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/655286a9-c261-43f2-ae31-80222cd3883b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "655286a9-c261-43f2-ae31-80222cd3883b" (UID: "655286a9-c261-43f2-ae31-80222cd3883b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.038460 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/655286a9-c261-43f2-ae31-80222cd3883b-config-data" (OuterVolumeSpecName: "config-data") pod "655286a9-c261-43f2-ae31-80222cd3883b" (UID: "655286a9-c261-43f2-ae31-80222cd3883b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.057452 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/655286a9-c261-43f2-ae31-80222cd3883b-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.057476 4813 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/655286a9-c261-43f2-ae31-80222cd3883b-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.057488 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/655286a9-c261-43f2-ae31-80222cd3883b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.057497 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/655286a9-c261-43f2-ae31-80222cd3883b-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.057505 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78n6q\" (UniqueName: \"kubernetes.io/projected/655286a9-c261-43f2-ae31-80222cd3883b-kube-api-access-78n6q\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.243927 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.264021 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.272704 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 06 16:03:54 crc kubenswrapper[4813]: E1206 16:03:54.273145 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="655286a9-c261-43f2-ae31-80222cd3883b" containerName="cinder-api" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.273162 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="655286a9-c261-43f2-ae31-80222cd3883b" containerName="cinder-api" Dec 06 16:03:54 crc kubenswrapper[4813]: E1206 16:03:54.273182 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="655286a9-c261-43f2-ae31-80222cd3883b" containerName="cinder-api-log" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.273189 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="655286a9-c261-43f2-ae31-80222cd3883b" containerName="cinder-api-log" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.273396 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="655286a9-c261-43f2-ae31-80222cd3883b" containerName="cinder-api" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.273417 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="655286a9-c261-43f2-ae31-80222cd3883b" containerName="cinder-api-log" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.274366 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.280402 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.282480 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.282542 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.285629 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.306023 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7c8fc98b5b-dwkdc" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.363990 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d-scripts\") pod \"cinder-api-0\" (UID: \"c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d\") " pod="openstack/cinder-api-0" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.364060 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d\") " pod="openstack/cinder-api-0" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.364227 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d-logs\") pod \"cinder-api-0\" (UID: \"c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d\") " pod="openstack/cinder-api-0" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.364275 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d\") " pod="openstack/cinder-api-0" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.364473 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d-config-data-custom\") pod \"cinder-api-0\" (UID: \"c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d\") " pod="openstack/cinder-api-0" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.364542 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d-public-tls-certs\") pod \"cinder-api-0\" (UID: \"c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d\") " pod="openstack/cinder-api-0" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.364564 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d-config-data\") pod \"cinder-api-0\" (UID: \"c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d\") " pod="openstack/cinder-api-0" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.364598 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d\") " pod="openstack/cinder-api-0" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.364646 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fzl7\" (UniqueName: \"kubernetes.io/projected/c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d-kube-api-access-7fzl7\") pod \"cinder-api-0\" (UID: \"c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d\") " pod="openstack/cinder-api-0" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.465415 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7c583806-afc7-4f95-8d4c-f0690157ca65-config\") pod \"7c583806-afc7-4f95-8d4c-f0690157ca65\" (UID: \"7c583806-afc7-4f95-8d4c-f0690157ca65\") " Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.465804 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42rxx\" (UniqueName: \"kubernetes.io/projected/7c583806-afc7-4f95-8d4c-f0690157ca65-kube-api-access-42rxx\") pod \"7c583806-afc7-4f95-8d4c-f0690157ca65\" (UID: \"7c583806-afc7-4f95-8d4c-f0690157ca65\") " Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.465974 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7c583806-afc7-4f95-8d4c-f0690157ca65-httpd-config\") pod \"7c583806-afc7-4f95-8d4c-f0690157ca65\" (UID: \"7c583806-afc7-4f95-8d4c-f0690157ca65\") " Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.466073 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c583806-afc7-4f95-8d4c-f0690157ca65-combined-ca-bundle\") pod \"7c583806-afc7-4f95-8d4c-f0690157ca65\" (UID: \"7c583806-afc7-4f95-8d4c-f0690157ca65\") " Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.466199 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c583806-afc7-4f95-8d4c-f0690157ca65-ovndb-tls-certs\") pod \"7c583806-afc7-4f95-8d4c-f0690157ca65\" (UID: \"7c583806-afc7-4f95-8d4c-f0690157ca65\") " Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.466474 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d-config-data-custom\") pod \"cinder-api-0\" (UID: \"c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d\") " pod="openstack/cinder-api-0" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.466575 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d-public-tls-certs\") pod \"cinder-api-0\" (UID: \"c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d\") " pod="openstack/cinder-api-0" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.466643 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d-config-data\") pod \"cinder-api-0\" (UID: \"c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d\") " pod="openstack/cinder-api-0" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.466723 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d\") " pod="openstack/cinder-api-0" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.466803 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fzl7\" (UniqueName: \"kubernetes.io/projected/c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d-kube-api-access-7fzl7\") pod \"cinder-api-0\" (UID: \"c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d\") " pod="openstack/cinder-api-0" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.466879 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d-scripts\") pod \"cinder-api-0\" (UID: \"c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d\") " pod="openstack/cinder-api-0" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.466968 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d\") " pod="openstack/cinder-api-0" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.467053 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d-logs\") pod \"cinder-api-0\" (UID: \"c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d\") " pod="openstack/cinder-api-0" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.467126 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d\") " pod="openstack/cinder-api-0" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.467571 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d-logs\") pod \"cinder-api-0\" (UID: \"c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d\") " pod="openstack/cinder-api-0" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.466891 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d\") " pod="openstack/cinder-api-0" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.474909 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d\") " pod="openstack/cinder-api-0" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.477527 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c583806-afc7-4f95-8d4c-f0690157ca65-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "7c583806-afc7-4f95-8d4c-f0690157ca65" (UID: "7c583806-afc7-4f95-8d4c-f0690157ca65"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.482594 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d\") " pod="openstack/cinder-api-0" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.482811 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fzl7\" (UniqueName: \"kubernetes.io/projected/c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d-kube-api-access-7fzl7\") pod \"cinder-api-0\" (UID: \"c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d\") " pod="openstack/cinder-api-0" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.493165 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d-public-tls-certs\") pod \"cinder-api-0\" (UID: \"c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d\") " pod="openstack/cinder-api-0" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.494758 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d-config-data-custom\") pod \"cinder-api-0\" (UID: \"c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d\") " pod="openstack/cinder-api-0" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.501278 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d-config-data\") pod \"cinder-api-0\" (UID: \"c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d\") " pod="openstack/cinder-api-0" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.507168 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d-scripts\") pod \"cinder-api-0\" (UID: \"c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d\") " pod="openstack/cinder-api-0" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.509247 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="655286a9-c261-43f2-ae31-80222cd3883b" path="/var/lib/kubelet/pods/655286a9-c261-43f2-ae31-80222cd3883b/volumes" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.510057 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f539e8ab-300c-426b-98e7-f9e87a6787b5" path="/var/lib/kubelet/pods/f539e8ab-300c-426b-98e7-f9e87a6787b5/volumes" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.510875 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c583806-afc7-4f95-8d4c-f0690157ca65-kube-api-access-42rxx" (OuterVolumeSpecName: "kube-api-access-42rxx") pod "7c583806-afc7-4f95-8d4c-f0690157ca65" (UID: "7c583806-afc7-4f95-8d4c-f0690157ca65"). InnerVolumeSpecName "kube-api-access-42rxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.568496 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42rxx\" (UniqueName: \"kubernetes.io/projected/7c583806-afc7-4f95-8d4c-f0690157ca65-kube-api-access-42rxx\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.568520 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7c583806-afc7-4f95-8d4c-f0690157ca65-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.573877 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c583806-afc7-4f95-8d4c-f0690157ca65-config" (OuterVolumeSpecName: "config") pod "7c583806-afc7-4f95-8d4c-f0690157ca65" (UID: "7c583806-afc7-4f95-8d4c-f0690157ca65"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.596584 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.604038 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c583806-afc7-4f95-8d4c-f0690157ca65-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7c583806-afc7-4f95-8d4c-f0690157ca65" (UID: "7c583806-afc7-4f95-8d4c-f0690157ca65"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.624176 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c583806-afc7-4f95-8d4c-f0690157ca65-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "7c583806-afc7-4f95-8d4c-f0690157ca65" (UID: "7c583806-afc7-4f95-8d4c-f0690157ca65"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.670116 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/7c583806-afc7-4f95-8d4c-f0690157ca65-config\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.670145 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c583806-afc7-4f95-8d4c-f0690157ca65-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.670155 4813 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c583806-afc7-4f95-8d4c-f0690157ca65-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.910445 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f56eb906-c881-4529-8c56-f11d6246a7f5","Type":"ContainerStarted","Data":"40dc9d35533a207fed9efd4b55d2b31ef37775793ed26251b96ffbc1b560e773"} Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.910759 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f56eb906-c881-4529-8c56-f11d6246a7f5","Type":"ContainerStarted","Data":"b85bb0beab9394e4bdeb1ef307e980902b5ef8c66785ef41533d7bd41ab8885f"} Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.913526 4813 generic.go:334] "Generic (PLEG): container finished" podID="7c583806-afc7-4f95-8d4c-f0690157ca65" containerID="9999291e2b8d7565b054e93fa369aa0a9392136209dae8df114b56c191ed251a" exitCode=0 Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.914517 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7c8fc98b5b-dwkdc" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.920597 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7c8fc98b5b-dwkdc" event={"ID":"7c583806-afc7-4f95-8d4c-f0690157ca65","Type":"ContainerDied","Data":"9999291e2b8d7565b054e93fa369aa0a9392136209dae8df114b56c191ed251a"} Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.920626 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7c8fc98b5b-dwkdc" event={"ID":"7c583806-afc7-4f95-8d4c-f0690157ca65","Type":"ContainerDied","Data":"877e6f39eec698c0411af6e3176db9a7abb1924e940b11e11f1627535bf4dd13"} Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.920641 4813 scope.go:117] "RemoveContainer" containerID="d6326e7e0b145e3ecd7b906844f31d37d84c0c433cced12074a74eecb1eb6d0d" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.950646 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7c8fc98b5b-dwkdc"] Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.963503 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7c8fc98b5b-dwkdc"] Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.964896 4813 scope.go:117] "RemoveContainer" containerID="9999291e2b8d7565b054e93fa369aa0a9392136209dae8df114b56c191ed251a" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.989580 4813 scope.go:117] "RemoveContainer" containerID="d6326e7e0b145e3ecd7b906844f31d37d84c0c433cced12074a74eecb1eb6d0d" Dec 06 16:03:54 crc kubenswrapper[4813]: E1206 16:03:54.990001 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6326e7e0b145e3ecd7b906844f31d37d84c0c433cced12074a74eecb1eb6d0d\": container with ID starting with d6326e7e0b145e3ecd7b906844f31d37d84c0c433cced12074a74eecb1eb6d0d not found: ID does not exist" containerID="d6326e7e0b145e3ecd7b906844f31d37d84c0c433cced12074a74eecb1eb6d0d" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.990037 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6326e7e0b145e3ecd7b906844f31d37d84c0c433cced12074a74eecb1eb6d0d"} err="failed to get container status \"d6326e7e0b145e3ecd7b906844f31d37d84c0c433cced12074a74eecb1eb6d0d\": rpc error: code = NotFound desc = could not find container \"d6326e7e0b145e3ecd7b906844f31d37d84c0c433cced12074a74eecb1eb6d0d\": container with ID starting with d6326e7e0b145e3ecd7b906844f31d37d84c0c433cced12074a74eecb1eb6d0d not found: ID does not exist" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.990064 4813 scope.go:117] "RemoveContainer" containerID="9999291e2b8d7565b054e93fa369aa0a9392136209dae8df114b56c191ed251a" Dec 06 16:03:54 crc kubenswrapper[4813]: E1206 16:03:54.990292 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9999291e2b8d7565b054e93fa369aa0a9392136209dae8df114b56c191ed251a\": container with ID starting with 9999291e2b8d7565b054e93fa369aa0a9392136209dae8df114b56c191ed251a not found: ID does not exist" containerID="9999291e2b8d7565b054e93fa369aa0a9392136209dae8df114b56c191ed251a" Dec 06 16:03:54 crc kubenswrapper[4813]: I1206 16:03:54.990313 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9999291e2b8d7565b054e93fa369aa0a9392136209dae8df114b56c191ed251a"} err="failed to get container status \"9999291e2b8d7565b054e93fa369aa0a9392136209dae8df114b56c191ed251a\": rpc error: code = NotFound desc = could not find container \"9999291e2b8d7565b054e93fa369aa0a9392136209dae8df114b56c191ed251a\": container with ID starting with 9999291e2b8d7565b054e93fa369aa0a9392136209dae8df114b56c191ed251a not found: ID does not exist" Dec 06 16:03:55 crc kubenswrapper[4813]: I1206 16:03:55.045793 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 06 16:03:55 crc kubenswrapper[4813]: W1206 16:03:55.054724 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc36fcd92_f5a1_4ec9_a1ee_fa6b54522e5d.slice/crio-7f530e08a5a2192d40e80a350e04fe13ab426226cf1d7fef6707d3e6b3def7e8 WatchSource:0}: Error finding container 7f530e08a5a2192d40e80a350e04fe13ab426226cf1d7fef6707d3e6b3def7e8: Status 404 returned error can't find the container with id 7f530e08a5a2192d40e80a350e04fe13ab426226cf1d7fef6707d3e6b3def7e8 Dec 06 16:03:55 crc kubenswrapper[4813]: I1206 16:03:55.714120 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-98dddfffb-l698f" Dec 06 16:03:55 crc kubenswrapper[4813]: I1206 16:03:55.932245 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f56eb906-c881-4529-8c56-f11d6246a7f5","Type":"ContainerStarted","Data":"a4cb16a6854b1fa60ec2178429bc0f364b1b0ecb0dd42db755d5cb6a6e3bb927"} Dec 06 16:03:55 crc kubenswrapper[4813]: I1206 16:03:55.934303 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d","Type":"ContainerStarted","Data":"d0edb6c521b0792e6ff312531939f2134b7d0f239055f254a44966c779d0e3c1"} Dec 06 16:03:55 crc kubenswrapper[4813]: I1206 16:03:55.934355 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d","Type":"ContainerStarted","Data":"7f530e08a5a2192d40e80a350e04fe13ab426226cf1d7fef6707d3e6b3def7e8"} Dec 06 16:03:56 crc kubenswrapper[4813]: I1206 16:03:56.497917 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c583806-afc7-4f95-8d4c-f0690157ca65" path="/var/lib/kubelet/pods/7c583806-afc7-4f95-8d4c-f0690157ca65/volumes" Dec 06 16:03:56 crc kubenswrapper[4813]: I1206 16:03:56.769782 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-98dddfffb-l698f" Dec 06 16:03:56 crc kubenswrapper[4813]: I1206 16:03:56.946987 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d","Type":"ContainerStarted","Data":"d1de678e5c1c9948198f8ee24f3fb69da60cb1bf962e75b3ca8760acec5b97f1"} Dec 06 16:03:56 crc kubenswrapper[4813]: I1206 16:03:56.947080 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 06 16:03:56 crc kubenswrapper[4813]: I1206 16:03:56.951714 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f56eb906-c881-4529-8c56-f11d6246a7f5","Type":"ContainerStarted","Data":"ac99e660696bd04d4805a5269756d79bf898a42e6018fd0e987bbaa07f8d7be8"} Dec 06 16:03:56 crc kubenswrapper[4813]: I1206 16:03:56.985819 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=2.985802223 podStartE2EDuration="2.985802223s" podCreationTimestamp="2025-12-06 16:03:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:03:56.976690273 +0000 UTC m=+1076.867569849" watchObservedRunningTime="2025-12-06 16:03:56.985802223 +0000 UTC m=+1076.876681799" Dec 06 16:03:57 crc kubenswrapper[4813]: I1206 16:03:57.961167 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f56eb906-c881-4529-8c56-f11d6246a7f5","Type":"ContainerStarted","Data":"2fe6951d33cbf1dba5a951aeaba29363bc6431914a943c66c24dc76a150e0c0b"} Dec 06 16:03:57 crc kubenswrapper[4813]: I1206 16:03:57.961435 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 16:03:57 crc kubenswrapper[4813]: I1206 16:03:57.978707 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.928568339 podStartE2EDuration="5.978689344s" podCreationTimestamp="2025-12-06 16:03:52 +0000 UTC" firstStartedPulling="2025-12-06 16:03:53.979216431 +0000 UTC m=+1073.870096007" lastFinishedPulling="2025-12-06 16:03:57.029337436 +0000 UTC m=+1076.920217012" observedRunningTime="2025-12-06 16:03:57.975905981 +0000 UTC m=+1077.866785567" watchObservedRunningTime="2025-12-06 16:03:57.978689344 +0000 UTC m=+1077.869568920" Dec 06 16:03:58 crc kubenswrapper[4813]: I1206 16:03:58.709540 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7fd6f7946b-kmm96" Dec 06 16:03:58 crc kubenswrapper[4813]: I1206 16:03:58.747893 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-bbc668d58-h8w67" Dec 06 16:03:58 crc kubenswrapper[4813]: I1206 16:03:58.971465 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6578955fd5-j49d2" Dec 06 16:03:59 crc kubenswrapper[4813]: I1206 16:03:59.045986 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-lnl2v"] Dec 06 16:03:59 crc kubenswrapper[4813]: I1206 16:03:59.046212 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" podUID="264bf858-a126-4005-9c98-0515834c4956" containerName="dnsmasq-dns" containerID="cri-o://399760fcc15809e2f6b6f950ca184113e25ea3d797744dab292c4dc1e9da448d" gracePeriod=10 Dec 06 16:03:59 crc kubenswrapper[4813]: I1206 16:03:59.246326 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 06 16:03:59 crc kubenswrapper[4813]: I1206 16:03:59.278128 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 16:03:59 crc kubenswrapper[4813]: I1206 16:03:59.581114 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" Dec 06 16:03:59 crc kubenswrapper[4813]: I1206 16:03:59.719712 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/264bf858-a126-4005-9c98-0515834c4956-dns-svc\") pod \"264bf858-a126-4005-9c98-0515834c4956\" (UID: \"264bf858-a126-4005-9c98-0515834c4956\") " Dec 06 16:03:59 crc kubenswrapper[4813]: I1206 16:03:59.719836 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/264bf858-a126-4005-9c98-0515834c4956-ovsdbserver-sb\") pod \"264bf858-a126-4005-9c98-0515834c4956\" (UID: \"264bf858-a126-4005-9c98-0515834c4956\") " Dec 06 16:03:59 crc kubenswrapper[4813]: I1206 16:03:59.719861 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/264bf858-a126-4005-9c98-0515834c4956-config\") pod \"264bf858-a126-4005-9c98-0515834c4956\" (UID: \"264bf858-a126-4005-9c98-0515834c4956\") " Dec 06 16:03:59 crc kubenswrapper[4813]: I1206 16:03:59.719931 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/264bf858-a126-4005-9c98-0515834c4956-dns-swift-storage-0\") pod \"264bf858-a126-4005-9c98-0515834c4956\" (UID: \"264bf858-a126-4005-9c98-0515834c4956\") " Dec 06 16:03:59 crc kubenswrapper[4813]: I1206 16:03:59.720071 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6796z\" (UniqueName: \"kubernetes.io/projected/264bf858-a126-4005-9c98-0515834c4956-kube-api-access-6796z\") pod \"264bf858-a126-4005-9c98-0515834c4956\" (UID: \"264bf858-a126-4005-9c98-0515834c4956\") " Dec 06 16:03:59 crc kubenswrapper[4813]: I1206 16:03:59.720107 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/264bf858-a126-4005-9c98-0515834c4956-ovsdbserver-nb\") pod \"264bf858-a126-4005-9c98-0515834c4956\" (UID: \"264bf858-a126-4005-9c98-0515834c4956\") " Dec 06 16:03:59 crc kubenswrapper[4813]: I1206 16:03:59.748441 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/264bf858-a126-4005-9c98-0515834c4956-kube-api-access-6796z" (OuterVolumeSpecName: "kube-api-access-6796z") pod "264bf858-a126-4005-9c98-0515834c4956" (UID: "264bf858-a126-4005-9c98-0515834c4956"). InnerVolumeSpecName "kube-api-access-6796z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:03:59 crc kubenswrapper[4813]: I1206 16:03:59.788625 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/264bf858-a126-4005-9c98-0515834c4956-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "264bf858-a126-4005-9c98-0515834c4956" (UID: "264bf858-a126-4005-9c98-0515834c4956"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:03:59 crc kubenswrapper[4813]: I1206 16:03:59.801542 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/264bf858-a126-4005-9c98-0515834c4956-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "264bf858-a126-4005-9c98-0515834c4956" (UID: "264bf858-a126-4005-9c98-0515834c4956"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:03:59 crc kubenswrapper[4813]: I1206 16:03:59.801773 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/264bf858-a126-4005-9c98-0515834c4956-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "264bf858-a126-4005-9c98-0515834c4956" (UID: "264bf858-a126-4005-9c98-0515834c4956"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:03:59 crc kubenswrapper[4813]: I1206 16:03:59.802592 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/264bf858-a126-4005-9c98-0515834c4956-config" (OuterVolumeSpecName: "config") pod "264bf858-a126-4005-9c98-0515834c4956" (UID: "264bf858-a126-4005-9c98-0515834c4956"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:03:59 crc kubenswrapper[4813]: I1206 16:03:59.813680 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/264bf858-a126-4005-9c98-0515834c4956-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "264bf858-a126-4005-9c98-0515834c4956" (UID: "264bf858-a126-4005-9c98-0515834c4956"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:03:59 crc kubenswrapper[4813]: I1206 16:03:59.822235 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/264bf858-a126-4005-9c98-0515834c4956-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:59 crc kubenswrapper[4813]: I1206 16:03:59.822294 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/264bf858-a126-4005-9c98-0515834c4956-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:59 crc kubenswrapper[4813]: I1206 16:03:59.822305 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/264bf858-a126-4005-9c98-0515834c4956-config\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:59 crc kubenswrapper[4813]: I1206 16:03:59.822314 4813 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/264bf858-a126-4005-9c98-0515834c4956-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:59 crc kubenswrapper[4813]: I1206 16:03:59.822324 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6796z\" (UniqueName: \"kubernetes.io/projected/264bf858-a126-4005-9c98-0515834c4956-kube-api-access-6796z\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:59 crc kubenswrapper[4813]: I1206 16:03:59.822332 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/264bf858-a126-4005-9c98-0515834c4956-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 16:03:59 crc kubenswrapper[4813]: I1206 16:03:59.977696 4813 generic.go:334] "Generic (PLEG): container finished" podID="264bf858-a126-4005-9c98-0515834c4956" containerID="399760fcc15809e2f6b6f950ca184113e25ea3d797744dab292c4dc1e9da448d" exitCode=0 Dec 06 16:03:59 crc kubenswrapper[4813]: I1206 16:03:59.977782 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" Dec 06 16:03:59 crc kubenswrapper[4813]: I1206 16:03:59.977795 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" event={"ID":"264bf858-a126-4005-9c98-0515834c4956","Type":"ContainerDied","Data":"399760fcc15809e2f6b6f950ca184113e25ea3d797744dab292c4dc1e9da448d"} Dec 06 16:03:59 crc kubenswrapper[4813]: I1206 16:03:59.977838 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-lnl2v" event={"ID":"264bf858-a126-4005-9c98-0515834c4956","Type":"ContainerDied","Data":"6cacbbb17ecd4d71a0f31e137350bebb3c7ae7742fcc9fa5a2350837b8e034d1"} Dec 06 16:03:59 crc kubenswrapper[4813]: I1206 16:03:59.977856 4813 scope.go:117] "RemoveContainer" containerID="399760fcc15809e2f6b6f950ca184113e25ea3d797744dab292c4dc1e9da448d" Dec 06 16:03:59 crc kubenswrapper[4813]: I1206 16:03:59.977925 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77" containerName="cinder-scheduler" containerID="cri-o://b3dad93bf75be6fdd022ed5d1d3a707a617523b4c6cbd751f73b7790327f0e38" gracePeriod=30 Dec 06 16:03:59 crc kubenswrapper[4813]: I1206 16:03:59.978047 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77" containerName="probe" containerID="cri-o://0eb7370c8ca2d86993b9b89e3aa1ac9a358a84d936bcf30ef9d764e36ca89976" gracePeriod=30 Dec 06 16:04:00 crc kubenswrapper[4813]: I1206 16:04:00.005989 4813 scope.go:117] "RemoveContainer" containerID="08467d4834e76df9230e68f5f05769d43c566c623bc6263be28d1c7a26156095" Dec 06 16:04:00 crc kubenswrapper[4813]: I1206 16:04:00.029324 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-lnl2v"] Dec 06 16:04:00 crc kubenswrapper[4813]: I1206 16:04:00.040249 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-lnl2v"] Dec 06 16:04:00 crc kubenswrapper[4813]: I1206 16:04:00.045213 4813 scope.go:117] "RemoveContainer" containerID="399760fcc15809e2f6b6f950ca184113e25ea3d797744dab292c4dc1e9da448d" Dec 06 16:04:00 crc kubenswrapper[4813]: E1206 16:04:00.046676 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"399760fcc15809e2f6b6f950ca184113e25ea3d797744dab292c4dc1e9da448d\": container with ID starting with 399760fcc15809e2f6b6f950ca184113e25ea3d797744dab292c4dc1e9da448d not found: ID does not exist" containerID="399760fcc15809e2f6b6f950ca184113e25ea3d797744dab292c4dc1e9da448d" Dec 06 16:04:00 crc kubenswrapper[4813]: I1206 16:04:00.046721 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"399760fcc15809e2f6b6f950ca184113e25ea3d797744dab292c4dc1e9da448d"} err="failed to get container status \"399760fcc15809e2f6b6f950ca184113e25ea3d797744dab292c4dc1e9da448d\": rpc error: code = NotFound desc = could not find container \"399760fcc15809e2f6b6f950ca184113e25ea3d797744dab292c4dc1e9da448d\": container with ID starting with 399760fcc15809e2f6b6f950ca184113e25ea3d797744dab292c4dc1e9da448d not found: ID does not exist" Dec 06 16:04:00 crc kubenswrapper[4813]: I1206 16:04:00.046754 4813 scope.go:117] "RemoveContainer" containerID="08467d4834e76df9230e68f5f05769d43c566c623bc6263be28d1c7a26156095" Dec 06 16:04:00 crc kubenswrapper[4813]: E1206 16:04:00.049621 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08467d4834e76df9230e68f5f05769d43c566c623bc6263be28d1c7a26156095\": container with ID starting with 08467d4834e76df9230e68f5f05769d43c566c623bc6263be28d1c7a26156095 not found: ID does not exist" containerID="08467d4834e76df9230e68f5f05769d43c566c623bc6263be28d1c7a26156095" Dec 06 16:04:00 crc kubenswrapper[4813]: I1206 16:04:00.049647 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08467d4834e76df9230e68f5f05769d43c566c623bc6263be28d1c7a26156095"} err="failed to get container status \"08467d4834e76df9230e68f5f05769d43c566c623bc6263be28d1c7a26156095\": rpc error: code = NotFound desc = could not find container \"08467d4834e76df9230e68f5f05769d43c566c623bc6263be28d1c7a26156095\": container with ID starting with 08467d4834e76df9230e68f5f05769d43c566c623bc6263be28d1c7a26156095 not found: ID does not exist" Dec 06 16:04:00 crc kubenswrapper[4813]: I1206 16:04:00.515034 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="264bf858-a126-4005-9c98-0515834c4956" path="/var/lib/kubelet/pods/264bf858-a126-4005-9c98-0515834c4956/volumes" Dec 06 16:04:00 crc kubenswrapper[4813]: I1206 16:04:00.884163 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7fd6f7946b-kmm96" Dec 06 16:04:00 crc kubenswrapper[4813]: I1206 16:04:00.939526 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-bbc668d58-h8w67"] Dec 06 16:04:00 crc kubenswrapper[4813]: I1206 16:04:00.939738 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-bbc668d58-h8w67" podUID="4bdc8329-162b-4592-b98e-ec7eda7f2ce1" containerName="horizon-log" containerID="cri-o://859333a385a1f24c723ad6a5689eae7998a7664c905e033301fe7108cfa531de" gracePeriod=30 Dec 06 16:04:00 crc kubenswrapper[4813]: I1206 16:04:00.940075 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-bbc668d58-h8w67" podUID="4bdc8329-162b-4592-b98e-ec7eda7f2ce1" containerName="horizon" containerID="cri-o://3307007b53464dbc4d2a32eb29a1577325f2d8df6457bc6adf713e100058d865" gracePeriod=30 Dec 06 16:04:00 crc kubenswrapper[4813]: I1206 16:04:00.958704 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-bbc668d58-h8w67" podUID="4bdc8329-162b-4592-b98e-ec7eda7f2ce1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Dec 06 16:04:01 crc kubenswrapper[4813]: I1206 16:04:01.999517 4813 generic.go:334] "Generic (PLEG): container finished" podID="b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77" containerID="0eb7370c8ca2d86993b9b89e3aa1ac9a358a84d936bcf30ef9d764e36ca89976" exitCode=0 Dec 06 16:04:01 crc kubenswrapper[4813]: I1206 16:04:01.999585 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77","Type":"ContainerDied","Data":"0eb7370c8ca2d86993b9b89e3aa1ac9a358a84d936bcf30ef9d764e36ca89976"} Dec 06 16:04:03 crc kubenswrapper[4813]: I1206 16:04:03.868846 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.000746 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-etc-machine-id\") pod \"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77\" (UID: \"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77\") " Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.000802 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-config-data\") pod \"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77\" (UID: \"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77\") " Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.000826 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-config-data-custom\") pod \"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77\" (UID: \"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77\") " Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.000878 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77" (UID: "b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.000901 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-scripts\") pod \"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77\" (UID: \"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77\") " Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.001030 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-combined-ca-bundle\") pod \"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77\" (UID: \"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77\") " Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.001083 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccfdp\" (UniqueName: \"kubernetes.io/projected/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-kube-api-access-ccfdp\") pod \"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77\" (UID: \"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77\") " Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.001406 4813 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.017031 4813 generic.go:334] "Generic (PLEG): container finished" podID="b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77" containerID="b3dad93bf75be6fdd022ed5d1d3a707a617523b4c6cbd751f73b7790327f0e38" exitCode=0 Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.017071 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77","Type":"ContainerDied","Data":"b3dad93bf75be6fdd022ed5d1d3a707a617523b4c6cbd751f73b7790327f0e38"} Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.017097 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77","Type":"ContainerDied","Data":"63fa3e880fd7361099030f8e2114673965cfc912c1bedbf25f3435a16306fc7d"} Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.017113 4813 scope.go:117] "RemoveContainer" containerID="0eb7370c8ca2d86993b9b89e3aa1ac9a358a84d936bcf30ef9d764e36ca89976" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.017231 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.025491 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-kube-api-access-ccfdp" (OuterVolumeSpecName: "kube-api-access-ccfdp") pod "b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77" (UID: "b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77"). InnerVolumeSpecName "kube-api-access-ccfdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.031359 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77" (UID: "b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.040498 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-scripts" (OuterVolumeSpecName: "scripts") pod "b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77" (UID: "b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.078461 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77" (UID: "b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.102606 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.102641 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ccfdp\" (UniqueName: \"kubernetes.io/projected/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-kube-api-access-ccfdp\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.102651 4813 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.102660 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.132140 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-config-data" (OuterVolumeSpecName: "config-data") pod "b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77" (UID: "b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.136016 4813 scope.go:117] "RemoveContainer" containerID="b3dad93bf75be6fdd022ed5d1d3a707a617523b4c6cbd751f73b7790327f0e38" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.198708 4813 scope.go:117] "RemoveContainer" containerID="0eb7370c8ca2d86993b9b89e3aa1ac9a358a84d936bcf30ef9d764e36ca89976" Dec 06 16:04:04 crc kubenswrapper[4813]: E1206 16:04:04.199145 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0eb7370c8ca2d86993b9b89e3aa1ac9a358a84d936bcf30ef9d764e36ca89976\": container with ID starting with 0eb7370c8ca2d86993b9b89e3aa1ac9a358a84d936bcf30ef9d764e36ca89976 not found: ID does not exist" containerID="0eb7370c8ca2d86993b9b89e3aa1ac9a358a84d936bcf30ef9d764e36ca89976" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.199170 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0eb7370c8ca2d86993b9b89e3aa1ac9a358a84d936bcf30ef9d764e36ca89976"} err="failed to get container status \"0eb7370c8ca2d86993b9b89e3aa1ac9a358a84d936bcf30ef9d764e36ca89976\": rpc error: code = NotFound desc = could not find container \"0eb7370c8ca2d86993b9b89e3aa1ac9a358a84d936bcf30ef9d764e36ca89976\": container with ID starting with 0eb7370c8ca2d86993b9b89e3aa1ac9a358a84d936bcf30ef9d764e36ca89976 not found: ID does not exist" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.199189 4813 scope.go:117] "RemoveContainer" containerID="b3dad93bf75be6fdd022ed5d1d3a707a617523b4c6cbd751f73b7790327f0e38" Dec 06 16:04:04 crc kubenswrapper[4813]: E1206 16:04:04.199419 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3dad93bf75be6fdd022ed5d1d3a707a617523b4c6cbd751f73b7790327f0e38\": container with ID starting with b3dad93bf75be6fdd022ed5d1d3a707a617523b4c6cbd751f73b7790327f0e38 not found: ID does not exist" containerID="b3dad93bf75be6fdd022ed5d1d3a707a617523b4c6cbd751f73b7790327f0e38" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.199435 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3dad93bf75be6fdd022ed5d1d3a707a617523b4c6cbd751f73b7790327f0e38"} err="failed to get container status \"b3dad93bf75be6fdd022ed5d1d3a707a617523b4c6cbd751f73b7790327f0e38\": rpc error: code = NotFound desc = could not find container \"b3dad93bf75be6fdd022ed5d1d3a707a617523b4c6cbd751f73b7790327f0e38\": container with ID starting with b3dad93bf75be6fdd022ed5d1d3a707a617523b4c6cbd751f73b7790327f0e38 not found: ID does not exist" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.204354 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.354272 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.361601 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.393108 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 16:04:04 crc kubenswrapper[4813]: E1206 16:04:04.393683 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77" containerName="probe" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.393752 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77" containerName="probe" Dec 06 16:04:04 crc kubenswrapper[4813]: E1206 16:04:04.393822 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c583806-afc7-4f95-8d4c-f0690157ca65" containerName="neutron-api" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.393871 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c583806-afc7-4f95-8d4c-f0690157ca65" containerName="neutron-api" Dec 06 16:04:04 crc kubenswrapper[4813]: E1206 16:04:04.393931 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="264bf858-a126-4005-9c98-0515834c4956" containerName="dnsmasq-dns" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.393987 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="264bf858-a126-4005-9c98-0515834c4956" containerName="dnsmasq-dns" Dec 06 16:04:04 crc kubenswrapper[4813]: E1206 16:04:04.394042 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c583806-afc7-4f95-8d4c-f0690157ca65" containerName="neutron-httpd" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.394090 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c583806-afc7-4f95-8d4c-f0690157ca65" containerName="neutron-httpd" Dec 06 16:04:04 crc kubenswrapper[4813]: E1206 16:04:04.394153 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77" containerName="cinder-scheduler" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.394211 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77" containerName="cinder-scheduler" Dec 06 16:04:04 crc kubenswrapper[4813]: E1206 16:04:04.394275 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="264bf858-a126-4005-9c98-0515834c4956" containerName="init" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.394325 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="264bf858-a126-4005-9c98-0515834c4956" containerName="init" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.394537 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c583806-afc7-4f95-8d4c-f0690157ca65" containerName="neutron-api" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.394601 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="264bf858-a126-4005-9c98-0515834c4956" containerName="dnsmasq-dns" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.394665 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77" containerName="probe" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.394730 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c583806-afc7-4f95-8d4c-f0690157ca65" containerName="neutron-httpd" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.394784 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77" containerName="cinder-scheduler" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.395687 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.400442 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.411906 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.496551 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77" path="/var/lib/kubelet/pods/b832f9b4-5cd5-4b5d-80ab-20cb2fcaad77/volumes" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.508941 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d707d75-8ac3-4256-9e9e-018e5caef917-scripts\") pod \"cinder-scheduler-0\" (UID: \"7d707d75-8ac3-4256-9e9e-018e5caef917\") " pod="openstack/cinder-scheduler-0" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.509014 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpvzb\" (UniqueName: \"kubernetes.io/projected/7d707d75-8ac3-4256-9e9e-018e5caef917-kube-api-access-tpvzb\") pod \"cinder-scheduler-0\" (UID: \"7d707d75-8ac3-4256-9e9e-018e5caef917\") " pod="openstack/cinder-scheduler-0" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.509039 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7d707d75-8ac3-4256-9e9e-018e5caef917-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7d707d75-8ac3-4256-9e9e-018e5caef917\") " pod="openstack/cinder-scheduler-0" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.509218 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d707d75-8ac3-4256-9e9e-018e5caef917-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7d707d75-8ac3-4256-9e9e-018e5caef917\") " pod="openstack/cinder-scheduler-0" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.509569 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d707d75-8ac3-4256-9e9e-018e5caef917-config-data\") pod \"cinder-scheduler-0\" (UID: \"7d707d75-8ac3-4256-9e9e-018e5caef917\") " pod="openstack/cinder-scheduler-0" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.509701 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7d707d75-8ac3-4256-9e9e-018e5caef917-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7d707d75-8ac3-4256-9e9e-018e5caef917\") " pod="openstack/cinder-scheduler-0" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.537674 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-bbc668d58-h8w67" podUID="4bdc8329-162b-4592-b98e-ec7eda7f2ce1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:57388->10.217.0.148:8443: read: connection reset by peer" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.611801 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d707d75-8ac3-4256-9e9e-018e5caef917-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7d707d75-8ac3-4256-9e9e-018e5caef917\") " pod="openstack/cinder-scheduler-0" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.611929 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d707d75-8ac3-4256-9e9e-018e5caef917-config-data\") pod \"cinder-scheduler-0\" (UID: \"7d707d75-8ac3-4256-9e9e-018e5caef917\") " pod="openstack/cinder-scheduler-0" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.611958 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7d707d75-8ac3-4256-9e9e-018e5caef917-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7d707d75-8ac3-4256-9e9e-018e5caef917\") " pod="openstack/cinder-scheduler-0" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.612009 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d707d75-8ac3-4256-9e9e-018e5caef917-scripts\") pod \"cinder-scheduler-0\" (UID: \"7d707d75-8ac3-4256-9e9e-018e5caef917\") " pod="openstack/cinder-scheduler-0" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.612051 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpvzb\" (UniqueName: \"kubernetes.io/projected/7d707d75-8ac3-4256-9e9e-018e5caef917-kube-api-access-tpvzb\") pod \"cinder-scheduler-0\" (UID: \"7d707d75-8ac3-4256-9e9e-018e5caef917\") " pod="openstack/cinder-scheduler-0" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.612094 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7d707d75-8ac3-4256-9e9e-018e5caef917-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7d707d75-8ac3-4256-9e9e-018e5caef917\") " pod="openstack/cinder-scheduler-0" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.612210 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7d707d75-8ac3-4256-9e9e-018e5caef917-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7d707d75-8ac3-4256-9e9e-018e5caef917\") " pod="openstack/cinder-scheduler-0" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.622209 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d707d75-8ac3-4256-9e9e-018e5caef917-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7d707d75-8ac3-4256-9e9e-018e5caef917\") " pod="openstack/cinder-scheduler-0" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.625067 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d707d75-8ac3-4256-9e9e-018e5caef917-scripts\") pod \"cinder-scheduler-0\" (UID: \"7d707d75-8ac3-4256-9e9e-018e5caef917\") " pod="openstack/cinder-scheduler-0" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.626488 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d707d75-8ac3-4256-9e9e-018e5caef917-config-data\") pod \"cinder-scheduler-0\" (UID: \"7d707d75-8ac3-4256-9e9e-018e5caef917\") " pod="openstack/cinder-scheduler-0" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.630224 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7d707d75-8ac3-4256-9e9e-018e5caef917-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7d707d75-8ac3-4256-9e9e-018e5caef917\") " pod="openstack/cinder-scheduler-0" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.657621 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpvzb\" (UniqueName: \"kubernetes.io/projected/7d707d75-8ac3-4256-9e9e-018e5caef917-kube-api-access-tpvzb\") pod \"cinder-scheduler-0\" (UID: \"7d707d75-8ac3-4256-9e9e-018e5caef917\") " pod="openstack/cinder-scheduler-0" Dec 06 16:04:04 crc kubenswrapper[4813]: I1206 16:04:04.711035 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 16:04:05 crc kubenswrapper[4813]: I1206 16:04:05.031777 4813 generic.go:334] "Generic (PLEG): container finished" podID="4bdc8329-162b-4592-b98e-ec7eda7f2ce1" containerID="3307007b53464dbc4d2a32eb29a1577325f2d8df6457bc6adf713e100058d865" exitCode=0 Dec 06 16:04:05 crc kubenswrapper[4813]: I1206 16:04:05.031818 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-bbc668d58-h8w67" event={"ID":"4bdc8329-162b-4592-b98e-ec7eda7f2ce1","Type":"ContainerDied","Data":"3307007b53464dbc4d2a32eb29a1577325f2d8df6457bc6adf713e100058d865"} Dec 06 16:04:05 crc kubenswrapper[4813]: I1206 16:04:05.237637 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 16:04:05 crc kubenswrapper[4813]: W1206 16:04:05.244297 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d707d75_8ac3_4256_9e9e_018e5caef917.slice/crio-29c0f330d5aa11a0eacd119bfe694b16677c6d4e72b09e095f4b3ef8e94a13bd WatchSource:0}: Error finding container 29c0f330d5aa11a0eacd119bfe694b16677c6d4e72b09e095f4b3ef8e94a13bd: Status 404 returned error can't find the container with id 29c0f330d5aa11a0eacd119bfe694b16677c6d4e72b09e095f4b3ef8e94a13bd Dec 06 16:04:05 crc kubenswrapper[4813]: I1206 16:04:05.761917 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-c4f8d8d9b-p5pqq" Dec 06 16:04:06 crc kubenswrapper[4813]: I1206 16:04:06.061813 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7d707d75-8ac3-4256-9e9e-018e5caef917","Type":"ContainerStarted","Data":"e67f4195d999f0ab112817f775341d8a0050fcfc7238f93c90d335652c30e069"} Dec 06 16:04:06 crc kubenswrapper[4813]: I1206 16:04:06.062074 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7d707d75-8ac3-4256-9e9e-018e5caef917","Type":"ContainerStarted","Data":"29c0f330d5aa11a0eacd119bfe694b16677c6d4e72b09e095f4b3ef8e94a13bd"} Dec 06 16:04:06 crc kubenswrapper[4813]: I1206 16:04:06.231893 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-bbc668d58-h8w67" podUID="4bdc8329-162b-4592-b98e-ec7eda7f2ce1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Dec 06 16:04:06 crc kubenswrapper[4813]: I1206 16:04:06.869785 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 06 16:04:07 crc kubenswrapper[4813]: I1206 16:04:07.070486 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7d707d75-8ac3-4256-9e9e-018e5caef917","Type":"ContainerStarted","Data":"b148630b846a82fbe47192b4bc267e55439fe084c26365703725d607fc05cc7d"} Dec 06 16:04:07 crc kubenswrapper[4813]: I1206 16:04:07.100392 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.100376167 podStartE2EDuration="3.100376167s" podCreationTimestamp="2025-12-06 16:04:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:04:07.094881793 +0000 UTC m=+1086.985761359" watchObservedRunningTime="2025-12-06 16:04:07.100376167 +0000 UTC m=+1086.991255743" Dec 06 16:04:09 crc kubenswrapper[4813]: I1206 16:04:09.712109 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 06 16:04:09 crc kubenswrapper[4813]: I1206 16:04:09.753559 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 06 16:04:09 crc kubenswrapper[4813]: I1206 16:04:09.754868 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 06 16:04:09 crc kubenswrapper[4813]: I1206 16:04:09.756700 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-p7gzk" Dec 06 16:04:09 crc kubenswrapper[4813]: I1206 16:04:09.757612 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 06 16:04:09 crc kubenswrapper[4813]: I1206 16:04:09.759352 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 06 16:04:09 crc kubenswrapper[4813]: I1206 16:04:09.766595 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 06 16:04:09 crc kubenswrapper[4813]: I1206 16:04:09.943604 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0447affb-1196-4c65-8e3d-936864d4fdba-combined-ca-bundle\") pod \"openstackclient\" (UID: \"0447affb-1196-4c65-8e3d-936864d4fdba\") " pod="openstack/openstackclient" Dec 06 16:04:09 crc kubenswrapper[4813]: I1206 16:04:09.943674 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqrdf\" (UniqueName: \"kubernetes.io/projected/0447affb-1196-4c65-8e3d-936864d4fdba-kube-api-access-gqrdf\") pod \"openstackclient\" (UID: \"0447affb-1196-4c65-8e3d-936864d4fdba\") " pod="openstack/openstackclient" Dec 06 16:04:09 crc kubenswrapper[4813]: I1206 16:04:09.943699 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0447affb-1196-4c65-8e3d-936864d4fdba-openstack-config\") pod \"openstackclient\" (UID: \"0447affb-1196-4c65-8e3d-936864d4fdba\") " pod="openstack/openstackclient" Dec 06 16:04:09 crc kubenswrapper[4813]: I1206 16:04:09.943771 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0447affb-1196-4c65-8e3d-936864d4fdba-openstack-config-secret\") pod \"openstackclient\" (UID: \"0447affb-1196-4c65-8e3d-936864d4fdba\") " pod="openstack/openstackclient" Dec 06 16:04:10 crc kubenswrapper[4813]: I1206 16:04:10.045359 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqrdf\" (UniqueName: \"kubernetes.io/projected/0447affb-1196-4c65-8e3d-936864d4fdba-kube-api-access-gqrdf\") pod \"openstackclient\" (UID: \"0447affb-1196-4c65-8e3d-936864d4fdba\") " pod="openstack/openstackclient" Dec 06 16:04:10 crc kubenswrapper[4813]: I1206 16:04:10.045424 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0447affb-1196-4c65-8e3d-936864d4fdba-openstack-config\") pod \"openstackclient\" (UID: \"0447affb-1196-4c65-8e3d-936864d4fdba\") " pod="openstack/openstackclient" Dec 06 16:04:10 crc kubenswrapper[4813]: I1206 16:04:10.045533 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0447affb-1196-4c65-8e3d-936864d4fdba-openstack-config-secret\") pod \"openstackclient\" (UID: \"0447affb-1196-4c65-8e3d-936864d4fdba\") " pod="openstack/openstackclient" Dec 06 16:04:10 crc kubenswrapper[4813]: I1206 16:04:10.045582 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0447affb-1196-4c65-8e3d-936864d4fdba-combined-ca-bundle\") pod \"openstackclient\" (UID: \"0447affb-1196-4c65-8e3d-936864d4fdba\") " pod="openstack/openstackclient" Dec 06 16:04:10 crc kubenswrapper[4813]: I1206 16:04:10.047086 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0447affb-1196-4c65-8e3d-936864d4fdba-openstack-config\") pod \"openstackclient\" (UID: \"0447affb-1196-4c65-8e3d-936864d4fdba\") " pod="openstack/openstackclient" Dec 06 16:04:10 crc kubenswrapper[4813]: I1206 16:04:10.053946 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0447affb-1196-4c65-8e3d-936864d4fdba-openstack-config-secret\") pod \"openstackclient\" (UID: \"0447affb-1196-4c65-8e3d-936864d4fdba\") " pod="openstack/openstackclient" Dec 06 16:04:10 crc kubenswrapper[4813]: I1206 16:04:10.054122 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0447affb-1196-4c65-8e3d-936864d4fdba-combined-ca-bundle\") pod \"openstackclient\" (UID: \"0447affb-1196-4c65-8e3d-936864d4fdba\") " pod="openstack/openstackclient" Dec 06 16:04:10 crc kubenswrapper[4813]: I1206 16:04:10.071517 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqrdf\" (UniqueName: \"kubernetes.io/projected/0447affb-1196-4c65-8e3d-936864d4fdba-kube-api-access-gqrdf\") pod \"openstackclient\" (UID: \"0447affb-1196-4c65-8e3d-936864d4fdba\") " pod="openstack/openstackclient" Dec 06 16:04:10 crc kubenswrapper[4813]: I1206 16:04:10.371953 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 06 16:04:10 crc kubenswrapper[4813]: I1206 16:04:10.818154 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 06 16:04:10 crc kubenswrapper[4813]: W1206 16:04:10.835432 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0447affb_1196_4c65_8e3d_936864d4fdba.slice/crio-44ee7fa22bb91260fdf0ebff457cc6479c424f6df6aae4338f12a4041a0ea68e WatchSource:0}: Error finding container 44ee7fa22bb91260fdf0ebff457cc6479c424f6df6aae4338f12a4041a0ea68e: Status 404 returned error can't find the container with id 44ee7fa22bb91260fdf0ebff457cc6479c424f6df6aae4338f12a4041a0ea68e Dec 06 16:04:11 crc kubenswrapper[4813]: I1206 16:04:11.099861 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"0447affb-1196-4c65-8e3d-936864d4fdba","Type":"ContainerStarted","Data":"44ee7fa22bb91260fdf0ebff457cc6479c424f6df6aae4338f12a4041a0ea68e"} Dec 06 16:04:11 crc kubenswrapper[4813]: I1206 16:04:11.969213 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-85cc9c6459-lnvlp"] Dec 06 16:04:11 crc kubenswrapper[4813]: I1206 16:04:11.970632 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-85cc9c6459-lnvlp" Dec 06 16:04:11 crc kubenswrapper[4813]: I1206 16:04:11.986427 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 06 16:04:11 crc kubenswrapper[4813]: I1206 16:04:11.986600 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 06 16:04:11 crc kubenswrapper[4813]: I1206 16:04:11.986719 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 06 16:04:11 crc kubenswrapper[4813]: I1206 16:04:11.992921 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-85cc9c6459-lnvlp"] Dec 06 16:04:12 crc kubenswrapper[4813]: I1206 16:04:12.080408 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4b18e9f-80a6-480e-878b-3acf0a5716a3-public-tls-certs\") pod \"swift-proxy-85cc9c6459-lnvlp\" (UID: \"c4b18e9f-80a6-480e-878b-3acf0a5716a3\") " pod="openstack/swift-proxy-85cc9c6459-lnvlp" Dec 06 16:04:12 crc kubenswrapper[4813]: I1206 16:04:12.080706 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4b18e9f-80a6-480e-878b-3acf0a5716a3-internal-tls-certs\") pod \"swift-proxy-85cc9c6459-lnvlp\" (UID: \"c4b18e9f-80a6-480e-878b-3acf0a5716a3\") " pod="openstack/swift-proxy-85cc9c6459-lnvlp" Dec 06 16:04:12 crc kubenswrapper[4813]: I1206 16:04:12.080735 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4b18e9f-80a6-480e-878b-3acf0a5716a3-config-data\") pod \"swift-proxy-85cc9c6459-lnvlp\" (UID: \"c4b18e9f-80a6-480e-878b-3acf0a5716a3\") " pod="openstack/swift-proxy-85cc9c6459-lnvlp" Dec 06 16:04:12 crc kubenswrapper[4813]: I1206 16:04:12.080751 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4b18e9f-80a6-480e-878b-3acf0a5716a3-log-httpd\") pod \"swift-proxy-85cc9c6459-lnvlp\" (UID: \"c4b18e9f-80a6-480e-878b-3acf0a5716a3\") " pod="openstack/swift-proxy-85cc9c6459-lnvlp" Dec 06 16:04:12 crc kubenswrapper[4813]: I1206 16:04:12.080797 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8tjj\" (UniqueName: \"kubernetes.io/projected/c4b18e9f-80a6-480e-878b-3acf0a5716a3-kube-api-access-l8tjj\") pod \"swift-proxy-85cc9c6459-lnvlp\" (UID: \"c4b18e9f-80a6-480e-878b-3acf0a5716a3\") " pod="openstack/swift-proxy-85cc9c6459-lnvlp" Dec 06 16:04:12 crc kubenswrapper[4813]: I1206 16:04:12.080861 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4b18e9f-80a6-480e-878b-3acf0a5716a3-combined-ca-bundle\") pod \"swift-proxy-85cc9c6459-lnvlp\" (UID: \"c4b18e9f-80a6-480e-878b-3acf0a5716a3\") " pod="openstack/swift-proxy-85cc9c6459-lnvlp" Dec 06 16:04:12 crc kubenswrapper[4813]: I1206 16:04:12.080886 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4b18e9f-80a6-480e-878b-3acf0a5716a3-run-httpd\") pod \"swift-proxy-85cc9c6459-lnvlp\" (UID: \"c4b18e9f-80a6-480e-878b-3acf0a5716a3\") " pod="openstack/swift-proxy-85cc9c6459-lnvlp" Dec 06 16:04:12 crc kubenswrapper[4813]: I1206 16:04:12.080919 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c4b18e9f-80a6-480e-878b-3acf0a5716a3-etc-swift\") pod \"swift-proxy-85cc9c6459-lnvlp\" (UID: \"c4b18e9f-80a6-480e-878b-3acf0a5716a3\") " pod="openstack/swift-proxy-85cc9c6459-lnvlp" Dec 06 16:04:12 crc kubenswrapper[4813]: I1206 16:04:12.182125 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4b18e9f-80a6-480e-878b-3acf0a5716a3-internal-tls-certs\") pod \"swift-proxy-85cc9c6459-lnvlp\" (UID: \"c4b18e9f-80a6-480e-878b-3acf0a5716a3\") " pod="openstack/swift-proxy-85cc9c6459-lnvlp" Dec 06 16:04:12 crc kubenswrapper[4813]: I1206 16:04:12.182171 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4b18e9f-80a6-480e-878b-3acf0a5716a3-config-data\") pod \"swift-proxy-85cc9c6459-lnvlp\" (UID: \"c4b18e9f-80a6-480e-878b-3acf0a5716a3\") " pod="openstack/swift-proxy-85cc9c6459-lnvlp" Dec 06 16:04:12 crc kubenswrapper[4813]: I1206 16:04:12.182191 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4b18e9f-80a6-480e-878b-3acf0a5716a3-log-httpd\") pod \"swift-proxy-85cc9c6459-lnvlp\" (UID: \"c4b18e9f-80a6-480e-878b-3acf0a5716a3\") " pod="openstack/swift-proxy-85cc9c6459-lnvlp" Dec 06 16:04:12 crc kubenswrapper[4813]: I1206 16:04:12.182244 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8tjj\" (UniqueName: \"kubernetes.io/projected/c4b18e9f-80a6-480e-878b-3acf0a5716a3-kube-api-access-l8tjj\") pod \"swift-proxy-85cc9c6459-lnvlp\" (UID: \"c4b18e9f-80a6-480e-878b-3acf0a5716a3\") " pod="openstack/swift-proxy-85cc9c6459-lnvlp" Dec 06 16:04:12 crc kubenswrapper[4813]: I1206 16:04:12.182294 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4b18e9f-80a6-480e-878b-3acf0a5716a3-combined-ca-bundle\") pod \"swift-proxy-85cc9c6459-lnvlp\" (UID: \"c4b18e9f-80a6-480e-878b-3acf0a5716a3\") " pod="openstack/swift-proxy-85cc9c6459-lnvlp" Dec 06 16:04:12 crc kubenswrapper[4813]: I1206 16:04:12.182318 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4b18e9f-80a6-480e-878b-3acf0a5716a3-run-httpd\") pod \"swift-proxy-85cc9c6459-lnvlp\" (UID: \"c4b18e9f-80a6-480e-878b-3acf0a5716a3\") " pod="openstack/swift-proxy-85cc9c6459-lnvlp" Dec 06 16:04:12 crc kubenswrapper[4813]: I1206 16:04:12.182353 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c4b18e9f-80a6-480e-878b-3acf0a5716a3-etc-swift\") pod \"swift-proxy-85cc9c6459-lnvlp\" (UID: \"c4b18e9f-80a6-480e-878b-3acf0a5716a3\") " pod="openstack/swift-proxy-85cc9c6459-lnvlp" Dec 06 16:04:12 crc kubenswrapper[4813]: I1206 16:04:12.182388 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4b18e9f-80a6-480e-878b-3acf0a5716a3-public-tls-certs\") pod \"swift-proxy-85cc9c6459-lnvlp\" (UID: \"c4b18e9f-80a6-480e-878b-3acf0a5716a3\") " pod="openstack/swift-proxy-85cc9c6459-lnvlp" Dec 06 16:04:12 crc kubenswrapper[4813]: I1206 16:04:12.184203 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4b18e9f-80a6-480e-878b-3acf0a5716a3-run-httpd\") pod \"swift-proxy-85cc9c6459-lnvlp\" (UID: \"c4b18e9f-80a6-480e-878b-3acf0a5716a3\") " pod="openstack/swift-proxy-85cc9c6459-lnvlp" Dec 06 16:04:12 crc kubenswrapper[4813]: I1206 16:04:12.184392 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4b18e9f-80a6-480e-878b-3acf0a5716a3-log-httpd\") pod \"swift-proxy-85cc9c6459-lnvlp\" (UID: \"c4b18e9f-80a6-480e-878b-3acf0a5716a3\") " pod="openstack/swift-proxy-85cc9c6459-lnvlp" Dec 06 16:04:12 crc kubenswrapper[4813]: I1206 16:04:12.189346 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4b18e9f-80a6-480e-878b-3acf0a5716a3-public-tls-certs\") pod \"swift-proxy-85cc9c6459-lnvlp\" (UID: \"c4b18e9f-80a6-480e-878b-3acf0a5716a3\") " pod="openstack/swift-proxy-85cc9c6459-lnvlp" Dec 06 16:04:12 crc kubenswrapper[4813]: I1206 16:04:12.189913 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4b18e9f-80a6-480e-878b-3acf0a5716a3-combined-ca-bundle\") pod \"swift-proxy-85cc9c6459-lnvlp\" (UID: \"c4b18e9f-80a6-480e-878b-3acf0a5716a3\") " pod="openstack/swift-proxy-85cc9c6459-lnvlp" Dec 06 16:04:12 crc kubenswrapper[4813]: I1206 16:04:12.198934 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4b18e9f-80a6-480e-878b-3acf0a5716a3-config-data\") pod \"swift-proxy-85cc9c6459-lnvlp\" (UID: \"c4b18e9f-80a6-480e-878b-3acf0a5716a3\") " pod="openstack/swift-proxy-85cc9c6459-lnvlp" Dec 06 16:04:12 crc kubenswrapper[4813]: I1206 16:04:12.200335 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c4b18e9f-80a6-480e-878b-3acf0a5716a3-etc-swift\") pod \"swift-proxy-85cc9c6459-lnvlp\" (UID: \"c4b18e9f-80a6-480e-878b-3acf0a5716a3\") " pod="openstack/swift-proxy-85cc9c6459-lnvlp" Dec 06 16:04:12 crc kubenswrapper[4813]: I1206 16:04:12.210889 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4b18e9f-80a6-480e-878b-3acf0a5716a3-internal-tls-certs\") pod \"swift-proxy-85cc9c6459-lnvlp\" (UID: \"c4b18e9f-80a6-480e-878b-3acf0a5716a3\") " pod="openstack/swift-proxy-85cc9c6459-lnvlp" Dec 06 16:04:12 crc kubenswrapper[4813]: I1206 16:04:12.211514 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8tjj\" (UniqueName: \"kubernetes.io/projected/c4b18e9f-80a6-480e-878b-3acf0a5716a3-kube-api-access-l8tjj\") pod \"swift-proxy-85cc9c6459-lnvlp\" (UID: \"c4b18e9f-80a6-480e-878b-3acf0a5716a3\") " pod="openstack/swift-proxy-85cc9c6459-lnvlp" Dec 06 16:04:12 crc kubenswrapper[4813]: I1206 16:04:12.293977 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-85cc9c6459-lnvlp" Dec 06 16:04:12 crc kubenswrapper[4813]: I1206 16:04:12.917222 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-85cc9c6459-lnvlp"] Dec 06 16:04:13 crc kubenswrapper[4813]: I1206 16:04:13.122469 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-85cc9c6459-lnvlp" event={"ID":"c4b18e9f-80a6-480e-878b-3acf0a5716a3","Type":"ContainerStarted","Data":"44bd737f3eb25b966d864779f8ac061ac77cbd4434da9ff06d85d71eae2ad99b"} Dec 06 16:04:14 crc kubenswrapper[4813]: I1206 16:04:14.135803 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-85cc9c6459-lnvlp" event={"ID":"c4b18e9f-80a6-480e-878b-3acf0a5716a3","Type":"ContainerStarted","Data":"9a0f4b67163ff88b58541a864e5ac41c27ccb1dc4e7852d2e9cfa8f4d77e5357"} Dec 06 16:04:14 crc kubenswrapper[4813]: I1206 16:04:14.136835 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-85cc9c6459-lnvlp" event={"ID":"c4b18e9f-80a6-480e-878b-3acf0a5716a3","Type":"ContainerStarted","Data":"ff98e6ce1d5bf3effc3094face794663961e48f59101bcff5e2cfcc9e155a825"} Dec 06 16:04:14 crc kubenswrapper[4813]: I1206 16:04:14.137016 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-85cc9c6459-lnvlp" Dec 06 16:04:14 crc kubenswrapper[4813]: I1206 16:04:14.137056 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-85cc9c6459-lnvlp" Dec 06 16:04:14 crc kubenswrapper[4813]: I1206 16:04:14.160073 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-85cc9c6459-lnvlp" podStartSLOduration=3.160060251 podStartE2EDuration="3.160060251s" podCreationTimestamp="2025-12-06 16:04:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:04:14.156126468 +0000 UTC m=+1094.047006044" watchObservedRunningTime="2025-12-06 16:04:14.160060251 +0000 UTC m=+1094.050939827" Dec 06 16:04:14 crc kubenswrapper[4813]: I1206 16:04:14.945663 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 06 16:04:16 crc kubenswrapper[4813]: I1206 16:04:16.151212 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:04:16 crc kubenswrapper[4813]: I1206 16:04:16.151705 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f56eb906-c881-4529-8c56-f11d6246a7f5" containerName="ceilometer-central-agent" containerID="cri-o://40dc9d35533a207fed9efd4b55d2b31ef37775793ed26251b96ffbc1b560e773" gracePeriod=30 Dec 06 16:04:16 crc kubenswrapper[4813]: I1206 16:04:16.152463 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f56eb906-c881-4529-8c56-f11d6246a7f5" containerName="proxy-httpd" containerID="cri-o://2fe6951d33cbf1dba5a951aeaba29363bc6431914a943c66c24dc76a150e0c0b" gracePeriod=30 Dec 06 16:04:16 crc kubenswrapper[4813]: I1206 16:04:16.152471 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f56eb906-c881-4529-8c56-f11d6246a7f5" containerName="ceilometer-notification-agent" containerID="cri-o://a4cb16a6854b1fa60ec2178429bc0f364b1b0ecb0dd42db755d5cb6a6e3bb927" gracePeriod=30 Dec 06 16:04:16 crc kubenswrapper[4813]: I1206 16:04:16.152472 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f56eb906-c881-4529-8c56-f11d6246a7f5" containerName="sg-core" containerID="cri-o://ac99e660696bd04d4805a5269756d79bf898a42e6018fd0e987bbaa07f8d7be8" gracePeriod=30 Dec 06 16:04:16 crc kubenswrapper[4813]: I1206 16:04:16.161452 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="f56eb906-c881-4529-8c56-f11d6246a7f5" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.166:3000/\": read tcp 10.217.0.2:38840->10.217.0.166:3000: read: connection reset by peer" Dec 06 16:04:16 crc kubenswrapper[4813]: I1206 16:04:16.231543 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-bbc668d58-h8w67" podUID="4bdc8329-162b-4592-b98e-ec7eda7f2ce1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Dec 06 16:04:17 crc kubenswrapper[4813]: I1206 16:04:17.169713 4813 generic.go:334] "Generic (PLEG): container finished" podID="f56eb906-c881-4529-8c56-f11d6246a7f5" containerID="2fe6951d33cbf1dba5a951aeaba29363bc6431914a943c66c24dc76a150e0c0b" exitCode=0 Dec 06 16:04:17 crc kubenswrapper[4813]: I1206 16:04:17.169748 4813 generic.go:334] "Generic (PLEG): container finished" podID="f56eb906-c881-4529-8c56-f11d6246a7f5" containerID="ac99e660696bd04d4805a5269756d79bf898a42e6018fd0e987bbaa07f8d7be8" exitCode=2 Dec 06 16:04:17 crc kubenswrapper[4813]: I1206 16:04:17.169758 4813 generic.go:334] "Generic (PLEG): container finished" podID="f56eb906-c881-4529-8c56-f11d6246a7f5" containerID="a4cb16a6854b1fa60ec2178429bc0f364b1b0ecb0dd42db755d5cb6a6e3bb927" exitCode=0 Dec 06 16:04:17 crc kubenswrapper[4813]: I1206 16:04:17.169767 4813 generic.go:334] "Generic (PLEG): container finished" podID="f56eb906-c881-4529-8c56-f11d6246a7f5" containerID="40dc9d35533a207fed9efd4b55d2b31ef37775793ed26251b96ffbc1b560e773" exitCode=0 Dec 06 16:04:17 crc kubenswrapper[4813]: I1206 16:04:17.169787 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f56eb906-c881-4529-8c56-f11d6246a7f5","Type":"ContainerDied","Data":"2fe6951d33cbf1dba5a951aeaba29363bc6431914a943c66c24dc76a150e0c0b"} Dec 06 16:04:17 crc kubenswrapper[4813]: I1206 16:04:17.169829 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f56eb906-c881-4529-8c56-f11d6246a7f5","Type":"ContainerDied","Data":"ac99e660696bd04d4805a5269756d79bf898a42e6018fd0e987bbaa07f8d7be8"} Dec 06 16:04:17 crc kubenswrapper[4813]: I1206 16:04:17.169839 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f56eb906-c881-4529-8c56-f11d6246a7f5","Type":"ContainerDied","Data":"a4cb16a6854b1fa60ec2178429bc0f364b1b0ecb0dd42db755d5cb6a6e3bb927"} Dec 06 16:04:17 crc kubenswrapper[4813]: I1206 16:04:17.169847 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f56eb906-c881-4529-8c56-f11d6246a7f5","Type":"ContainerDied","Data":"40dc9d35533a207fed9efd4b55d2b31ef37775793ed26251b96ffbc1b560e773"} Dec 06 16:04:21 crc kubenswrapper[4813]: I1206 16:04:21.904565 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.080964 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kkfc\" (UniqueName: \"kubernetes.io/projected/f56eb906-c881-4529-8c56-f11d6246a7f5-kube-api-access-7kkfc\") pod \"f56eb906-c881-4529-8c56-f11d6246a7f5\" (UID: \"f56eb906-c881-4529-8c56-f11d6246a7f5\") " Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.081066 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f56eb906-c881-4529-8c56-f11d6246a7f5-combined-ca-bundle\") pod \"f56eb906-c881-4529-8c56-f11d6246a7f5\" (UID: \"f56eb906-c881-4529-8c56-f11d6246a7f5\") " Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.081158 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f56eb906-c881-4529-8c56-f11d6246a7f5-scripts\") pod \"f56eb906-c881-4529-8c56-f11d6246a7f5\" (UID: \"f56eb906-c881-4529-8c56-f11d6246a7f5\") " Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.081187 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f56eb906-c881-4529-8c56-f11d6246a7f5-config-data\") pod \"f56eb906-c881-4529-8c56-f11d6246a7f5\" (UID: \"f56eb906-c881-4529-8c56-f11d6246a7f5\") " Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.081226 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f56eb906-c881-4529-8c56-f11d6246a7f5-sg-core-conf-yaml\") pod \"f56eb906-c881-4529-8c56-f11d6246a7f5\" (UID: \"f56eb906-c881-4529-8c56-f11d6246a7f5\") " Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.081282 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f56eb906-c881-4529-8c56-f11d6246a7f5-run-httpd\") pod \"f56eb906-c881-4529-8c56-f11d6246a7f5\" (UID: \"f56eb906-c881-4529-8c56-f11d6246a7f5\") " Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.081306 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f56eb906-c881-4529-8c56-f11d6246a7f5-log-httpd\") pod \"f56eb906-c881-4529-8c56-f11d6246a7f5\" (UID: \"f56eb906-c881-4529-8c56-f11d6246a7f5\") " Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.081975 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f56eb906-c881-4529-8c56-f11d6246a7f5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f56eb906-c881-4529-8c56-f11d6246a7f5" (UID: "f56eb906-c881-4529-8c56-f11d6246a7f5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.083190 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f56eb906-c881-4529-8c56-f11d6246a7f5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f56eb906-c881-4529-8c56-f11d6246a7f5" (UID: "f56eb906-c881-4529-8c56-f11d6246a7f5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.086067 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f56eb906-c881-4529-8c56-f11d6246a7f5-kube-api-access-7kkfc" (OuterVolumeSpecName: "kube-api-access-7kkfc") pod "f56eb906-c881-4529-8c56-f11d6246a7f5" (UID: "f56eb906-c881-4529-8c56-f11d6246a7f5"). InnerVolumeSpecName "kube-api-access-7kkfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.087058 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f56eb906-c881-4529-8c56-f11d6246a7f5-scripts" (OuterVolumeSpecName: "scripts") pod "f56eb906-c881-4529-8c56-f11d6246a7f5" (UID: "f56eb906-c881-4529-8c56-f11d6246a7f5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.122017 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f56eb906-c881-4529-8c56-f11d6246a7f5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f56eb906-c881-4529-8c56-f11d6246a7f5" (UID: "f56eb906-c881-4529-8c56-f11d6246a7f5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.212686 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f56eb906-c881-4529-8c56-f11d6246a7f5-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.212715 4813 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f56eb906-c881-4529-8c56-f11d6246a7f5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.212725 4813 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f56eb906-c881-4529-8c56-f11d6246a7f5-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.212734 4813 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f56eb906-c881-4529-8c56-f11d6246a7f5-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.212742 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kkfc\" (UniqueName: \"kubernetes.io/projected/f56eb906-c881-4529-8c56-f11d6246a7f5-kube-api-access-7kkfc\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.216975 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f56eb906-c881-4529-8c56-f11d6246a7f5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f56eb906-c881-4529-8c56-f11d6246a7f5" (UID: "f56eb906-c881-4529-8c56-f11d6246a7f5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.217124 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f56eb906-c881-4529-8c56-f11d6246a7f5-config-data" (OuterVolumeSpecName: "config-data") pod "f56eb906-c881-4529-8c56-f11d6246a7f5" (UID: "f56eb906-c881-4529-8c56-f11d6246a7f5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.224515 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"0447affb-1196-4c65-8e3d-936864d4fdba","Type":"ContainerStarted","Data":"0a360173d66d9a36bb8fa32b28e8508755d3def356c6bc0af3d298b9eccb990c"} Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.228162 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f56eb906-c881-4529-8c56-f11d6246a7f5","Type":"ContainerDied","Data":"b85bb0beab9394e4bdeb1ef307e980902b5ef8c66785ef41533d7bd41ab8885f"} Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.228210 4813 scope.go:117] "RemoveContainer" containerID="2fe6951d33cbf1dba5a951aeaba29363bc6431914a943c66c24dc76a150e0c0b" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.228344 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.246502 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.477892133 podStartE2EDuration="13.246487838s" podCreationTimestamp="2025-12-06 16:04:09 +0000 UTC" firstStartedPulling="2025-12-06 16:04:10.837011447 +0000 UTC m=+1090.727891023" lastFinishedPulling="2025-12-06 16:04:21.605607152 +0000 UTC m=+1101.496486728" observedRunningTime="2025-12-06 16:04:22.240511861 +0000 UTC m=+1102.131391437" watchObservedRunningTime="2025-12-06 16:04:22.246487838 +0000 UTC m=+1102.137367414" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.268219 4813 scope.go:117] "RemoveContainer" containerID="ac99e660696bd04d4805a5269756d79bf898a42e6018fd0e987bbaa07f8d7be8" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.279017 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.300444 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.310877 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:04:22 crc kubenswrapper[4813]: E1206 16:04:22.311295 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f56eb906-c881-4529-8c56-f11d6246a7f5" containerName="proxy-httpd" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.311310 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f56eb906-c881-4529-8c56-f11d6246a7f5" containerName="proxy-httpd" Dec 06 16:04:22 crc kubenswrapper[4813]: E1206 16:04:22.311323 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f56eb906-c881-4529-8c56-f11d6246a7f5" containerName="sg-core" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.311328 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f56eb906-c881-4529-8c56-f11d6246a7f5" containerName="sg-core" Dec 06 16:04:22 crc kubenswrapper[4813]: E1206 16:04:22.311348 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f56eb906-c881-4529-8c56-f11d6246a7f5" containerName="ceilometer-notification-agent" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.311354 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f56eb906-c881-4529-8c56-f11d6246a7f5" containerName="ceilometer-notification-agent" Dec 06 16:04:22 crc kubenswrapper[4813]: E1206 16:04:22.311366 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f56eb906-c881-4529-8c56-f11d6246a7f5" containerName="ceilometer-central-agent" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.311382 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f56eb906-c881-4529-8c56-f11d6246a7f5" containerName="ceilometer-central-agent" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.311585 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f56eb906-c881-4529-8c56-f11d6246a7f5" containerName="proxy-httpd" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.311602 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f56eb906-c881-4529-8c56-f11d6246a7f5" containerName="sg-core" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.311608 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f56eb906-c881-4529-8c56-f11d6246a7f5" containerName="ceilometer-central-agent" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.311618 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f56eb906-c881-4529-8c56-f11d6246a7f5" containerName="ceilometer-notification-agent" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.313230 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-85cc9c6459-lnvlp" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.313270 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-85cc9c6459-lnvlp" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.314083 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.314177 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f56eb906-c881-4529-8c56-f11d6246a7f5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.314197 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f56eb906-c881-4529-8c56-f11d6246a7f5-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.318090 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.319716 4813 scope.go:117] "RemoveContainer" containerID="a4cb16a6854b1fa60ec2178429bc0f364b1b0ecb0dd42db755d5cb6a6e3bb927" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.320275 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.321180 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.395693 4813 scope.go:117] "RemoveContainer" containerID="40dc9d35533a207fed9efd4b55d2b31ef37775793ed26251b96ffbc1b560e773" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.415153 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54c36b42-5bd6-4c47-92c2-80ac1be6646f-scripts\") pod \"ceilometer-0\" (UID: \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\") " pod="openstack/ceilometer-0" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.415230 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54c36b42-5bd6-4c47-92c2-80ac1be6646f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\") " pod="openstack/ceilometer-0" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.415318 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sjw5\" (UniqueName: \"kubernetes.io/projected/54c36b42-5bd6-4c47-92c2-80ac1be6646f-kube-api-access-9sjw5\") pod \"ceilometer-0\" (UID: \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\") " pod="openstack/ceilometer-0" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.415339 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54c36b42-5bd6-4c47-92c2-80ac1be6646f-log-httpd\") pod \"ceilometer-0\" (UID: \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\") " pod="openstack/ceilometer-0" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.415353 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/54c36b42-5bd6-4c47-92c2-80ac1be6646f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\") " pod="openstack/ceilometer-0" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.415390 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54c36b42-5bd6-4c47-92c2-80ac1be6646f-config-data\") pod \"ceilometer-0\" (UID: \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\") " pod="openstack/ceilometer-0" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.415450 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54c36b42-5bd6-4c47-92c2-80ac1be6646f-run-httpd\") pod \"ceilometer-0\" (UID: \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\") " pod="openstack/ceilometer-0" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.504986 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f56eb906-c881-4529-8c56-f11d6246a7f5" path="/var/lib/kubelet/pods/f56eb906-c881-4529-8c56-f11d6246a7f5/volumes" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.517496 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54c36b42-5bd6-4c47-92c2-80ac1be6646f-run-httpd\") pod \"ceilometer-0\" (UID: \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\") " pod="openstack/ceilometer-0" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.517554 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54c36b42-5bd6-4c47-92c2-80ac1be6646f-scripts\") pod \"ceilometer-0\" (UID: \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\") " pod="openstack/ceilometer-0" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.517595 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54c36b42-5bd6-4c47-92c2-80ac1be6646f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\") " pod="openstack/ceilometer-0" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.517640 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sjw5\" (UniqueName: \"kubernetes.io/projected/54c36b42-5bd6-4c47-92c2-80ac1be6646f-kube-api-access-9sjw5\") pod \"ceilometer-0\" (UID: \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\") " pod="openstack/ceilometer-0" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.517661 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54c36b42-5bd6-4c47-92c2-80ac1be6646f-log-httpd\") pod \"ceilometer-0\" (UID: \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\") " pod="openstack/ceilometer-0" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.517678 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/54c36b42-5bd6-4c47-92c2-80ac1be6646f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\") " pod="openstack/ceilometer-0" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.517822 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54c36b42-5bd6-4c47-92c2-80ac1be6646f-config-data\") pod \"ceilometer-0\" (UID: \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\") " pod="openstack/ceilometer-0" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.518058 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54c36b42-5bd6-4c47-92c2-80ac1be6646f-run-httpd\") pod \"ceilometer-0\" (UID: \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\") " pod="openstack/ceilometer-0" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.524275 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54c36b42-5bd6-4c47-92c2-80ac1be6646f-log-httpd\") pod \"ceilometer-0\" (UID: \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\") " pod="openstack/ceilometer-0" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.525462 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54c36b42-5bd6-4c47-92c2-80ac1be6646f-scripts\") pod \"ceilometer-0\" (UID: \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\") " pod="openstack/ceilometer-0" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.525600 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54c36b42-5bd6-4c47-92c2-80ac1be6646f-config-data\") pod \"ceilometer-0\" (UID: \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\") " pod="openstack/ceilometer-0" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.532318 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54c36b42-5bd6-4c47-92c2-80ac1be6646f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\") " pod="openstack/ceilometer-0" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.532941 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/54c36b42-5bd6-4c47-92c2-80ac1be6646f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\") " pod="openstack/ceilometer-0" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.546121 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sjw5\" (UniqueName: \"kubernetes.io/projected/54c36b42-5bd6-4c47-92c2-80ac1be6646f-kube-api-access-9sjw5\") pod \"ceilometer-0\" (UID: \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\") " pod="openstack/ceilometer-0" Dec 06 16:04:22 crc kubenswrapper[4813]: I1206 16:04:22.673995 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 16:04:23 crc kubenswrapper[4813]: I1206 16:04:23.180053 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:04:23 crc kubenswrapper[4813]: I1206 16:04:23.241119 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54c36b42-5bd6-4c47-92c2-80ac1be6646f","Type":"ContainerStarted","Data":"1ad6ab4359e25a54c11e3c95cf56e03b248766e5faac36d0faad3c482fef4ee3"} Dec 06 16:04:24 crc kubenswrapper[4813]: I1206 16:04:24.245760 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:04:24 crc kubenswrapper[4813]: I1206 16:04:24.260182 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54c36b42-5bd6-4c47-92c2-80ac1be6646f","Type":"ContainerStarted","Data":"cd8d1a683279343806f365da7844e6ec389a6e50ac444b8e9f03178e55a3f5da"} Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.081385 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-mxrr6"] Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.082830 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-mxrr6" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.110436 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-mxrr6"] Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.164962 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a06cadf5-56f1-439f-8deb-2aaedba66c8a-operator-scripts\") pod \"nova-api-db-create-mxrr6\" (UID: \"a06cadf5-56f1-439f-8deb-2aaedba66c8a\") " pod="openstack/nova-api-db-create-mxrr6" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.165055 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsrtv\" (UniqueName: \"kubernetes.io/projected/a06cadf5-56f1-439f-8deb-2aaedba66c8a-kube-api-access-qsrtv\") pod \"nova-api-db-create-mxrr6\" (UID: \"a06cadf5-56f1-439f-8deb-2aaedba66c8a\") " pod="openstack/nova-api-db-create-mxrr6" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.181101 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-t86b9"] Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.188517 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-t86b9" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.203838 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-t86b9"] Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.236945 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-3fbf-account-create-update-ksdbr"] Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.238127 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3fbf-account-create-update-ksdbr" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.241636 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.259993 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-3fbf-account-create-update-ksdbr"] Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.267222 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsrtv\" (UniqueName: \"kubernetes.io/projected/a06cadf5-56f1-439f-8deb-2aaedba66c8a-kube-api-access-qsrtv\") pod \"nova-api-db-create-mxrr6\" (UID: \"a06cadf5-56f1-439f-8deb-2aaedba66c8a\") " pod="openstack/nova-api-db-create-mxrr6" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.267312 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47a481f6-b873-4920-a273-6e65f8d6503a-operator-scripts\") pod \"nova-cell0-db-create-t86b9\" (UID: \"47a481f6-b873-4920-a273-6e65f8d6503a\") " pod="openstack/nova-cell0-db-create-t86b9" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.267348 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdh5n\" (UniqueName: \"kubernetes.io/projected/47a481f6-b873-4920-a273-6e65f8d6503a-kube-api-access-vdh5n\") pod \"nova-cell0-db-create-t86b9\" (UID: \"47a481f6-b873-4920-a273-6e65f8d6503a\") " pod="openstack/nova-cell0-db-create-t86b9" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.267394 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a06cadf5-56f1-439f-8deb-2aaedba66c8a-operator-scripts\") pod \"nova-api-db-create-mxrr6\" (UID: \"a06cadf5-56f1-439f-8deb-2aaedba66c8a\") " pod="openstack/nova-api-db-create-mxrr6" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.268038 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a06cadf5-56f1-439f-8deb-2aaedba66c8a-operator-scripts\") pod \"nova-api-db-create-mxrr6\" (UID: \"a06cadf5-56f1-439f-8deb-2aaedba66c8a\") " pod="openstack/nova-api-db-create-mxrr6" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.279901 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54c36b42-5bd6-4c47-92c2-80ac1be6646f","Type":"ContainerStarted","Data":"5de8e29fd04a40c2e172e5c514b2c0058f683d1db695f38c26b1ff48816a5908"} Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.279940 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54c36b42-5bd6-4c47-92c2-80ac1be6646f","Type":"ContainerStarted","Data":"f59deba9a17c1dd00efa199a4e897a0ab13ac64172156ed7d55d30bf2e48f493"} Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.300907 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsrtv\" (UniqueName: \"kubernetes.io/projected/a06cadf5-56f1-439f-8deb-2aaedba66c8a-kube-api-access-qsrtv\") pod \"nova-api-db-create-mxrr6\" (UID: \"a06cadf5-56f1-439f-8deb-2aaedba66c8a\") " pod="openstack/nova-api-db-create-mxrr6" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.368943 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47a481f6-b873-4920-a273-6e65f8d6503a-operator-scripts\") pod \"nova-cell0-db-create-t86b9\" (UID: \"47a481f6-b873-4920-a273-6e65f8d6503a\") " pod="openstack/nova-cell0-db-create-t86b9" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.369028 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdh5n\" (UniqueName: \"kubernetes.io/projected/47a481f6-b873-4920-a273-6e65f8d6503a-kube-api-access-vdh5n\") pod \"nova-cell0-db-create-t86b9\" (UID: \"47a481f6-b873-4920-a273-6e65f8d6503a\") " pod="openstack/nova-cell0-db-create-t86b9" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.369075 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghjhf\" (UniqueName: \"kubernetes.io/projected/dc35c201-8f5a-418b-a615-c04dff409d8d-kube-api-access-ghjhf\") pod \"nova-api-3fbf-account-create-update-ksdbr\" (UID: \"dc35c201-8f5a-418b-a615-c04dff409d8d\") " pod="openstack/nova-api-3fbf-account-create-update-ksdbr" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.369098 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc35c201-8f5a-418b-a615-c04dff409d8d-operator-scripts\") pod \"nova-api-3fbf-account-create-update-ksdbr\" (UID: \"dc35c201-8f5a-418b-a615-c04dff409d8d\") " pod="openstack/nova-api-3fbf-account-create-update-ksdbr" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.372946 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47a481f6-b873-4920-a273-6e65f8d6503a-operator-scripts\") pod \"nova-cell0-db-create-t86b9\" (UID: \"47a481f6-b873-4920-a273-6e65f8d6503a\") " pod="openstack/nova-cell0-db-create-t86b9" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.398692 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-mxrr6" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.407875 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdh5n\" (UniqueName: \"kubernetes.io/projected/47a481f6-b873-4920-a273-6e65f8d6503a-kube-api-access-vdh5n\") pod \"nova-cell0-db-create-t86b9\" (UID: \"47a481f6-b873-4920-a273-6e65f8d6503a\") " pod="openstack/nova-cell0-db-create-t86b9" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.430248 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-r2fvl"] Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.431465 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-r2fvl" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.442974 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-r2fvl"] Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.471001 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghjhf\" (UniqueName: \"kubernetes.io/projected/dc35c201-8f5a-418b-a615-c04dff409d8d-kube-api-access-ghjhf\") pod \"nova-api-3fbf-account-create-update-ksdbr\" (UID: \"dc35c201-8f5a-418b-a615-c04dff409d8d\") " pod="openstack/nova-api-3fbf-account-create-update-ksdbr" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.471052 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc35c201-8f5a-418b-a615-c04dff409d8d-operator-scripts\") pod \"nova-api-3fbf-account-create-update-ksdbr\" (UID: \"dc35c201-8f5a-418b-a615-c04dff409d8d\") " pod="openstack/nova-api-3fbf-account-create-update-ksdbr" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.471910 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc35c201-8f5a-418b-a615-c04dff409d8d-operator-scripts\") pod \"nova-api-3fbf-account-create-update-ksdbr\" (UID: \"dc35c201-8f5a-418b-a615-c04dff409d8d\") " pod="openstack/nova-api-3fbf-account-create-update-ksdbr" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.509168 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghjhf\" (UniqueName: \"kubernetes.io/projected/dc35c201-8f5a-418b-a615-c04dff409d8d-kube-api-access-ghjhf\") pod \"nova-api-3fbf-account-create-update-ksdbr\" (UID: \"dc35c201-8f5a-418b-a615-c04dff409d8d\") " pod="openstack/nova-api-3fbf-account-create-update-ksdbr" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.509707 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-t86b9" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.517905 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-56b5-account-create-update-98zzg"] Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.518977 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-56b5-account-create-update-98zzg" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.529190 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.531031 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-56b5-account-create-update-98zzg"] Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.560930 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3fbf-account-create-update-ksdbr" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.574412 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/532ac93c-0c2c-4bca-b9f2-e4f066913dbf-operator-scripts\") pod \"nova-cell1-db-create-r2fvl\" (UID: \"532ac93c-0c2c-4bca-b9f2-e4f066913dbf\") " pod="openstack/nova-cell1-db-create-r2fvl" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.574842 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glq8s\" (UniqueName: \"kubernetes.io/projected/532ac93c-0c2c-4bca-b9f2-e4f066913dbf-kube-api-access-glq8s\") pod \"nova-cell1-db-create-r2fvl\" (UID: \"532ac93c-0c2c-4bca-b9f2-e4f066913dbf\") " pod="openstack/nova-cell1-db-create-r2fvl" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.600653 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db9a-account-create-update-kb927"] Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.601785 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db9a-account-create-update-kb927" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.606086 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.610189 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db9a-account-create-update-kb927"] Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.679352 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4p77\" (UniqueName: \"kubernetes.io/projected/5d6f9192-b64d-4390-a4bd-4ff0a33bc31b-kube-api-access-s4p77\") pod \"nova-cell0-56b5-account-create-update-98zzg\" (UID: \"5d6f9192-b64d-4390-a4bd-4ff0a33bc31b\") " pod="openstack/nova-cell0-56b5-account-create-update-98zzg" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.679455 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl84k\" (UniqueName: \"kubernetes.io/projected/7fbe6119-47a9-4188-b8c0-a9e87d8ea09f-kube-api-access-tl84k\") pod \"nova-cell1-db9a-account-create-update-kb927\" (UID: \"7fbe6119-47a9-4188-b8c0-a9e87d8ea09f\") " pod="openstack/nova-cell1-db9a-account-create-update-kb927" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.683137 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/532ac93c-0c2c-4bca-b9f2-e4f066913dbf-operator-scripts\") pod \"nova-cell1-db-create-r2fvl\" (UID: \"532ac93c-0c2c-4bca-b9f2-e4f066913dbf\") " pod="openstack/nova-cell1-db-create-r2fvl" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.683187 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5d6f9192-b64d-4390-a4bd-4ff0a33bc31b-operator-scripts\") pod \"nova-cell0-56b5-account-create-update-98zzg\" (UID: \"5d6f9192-b64d-4390-a4bd-4ff0a33bc31b\") " pod="openstack/nova-cell0-56b5-account-create-update-98zzg" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.683222 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7fbe6119-47a9-4188-b8c0-a9e87d8ea09f-operator-scripts\") pod \"nova-cell1-db9a-account-create-update-kb927\" (UID: \"7fbe6119-47a9-4188-b8c0-a9e87d8ea09f\") " pod="openstack/nova-cell1-db9a-account-create-update-kb927" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.684831 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/532ac93c-0c2c-4bca-b9f2-e4f066913dbf-operator-scripts\") pod \"nova-cell1-db-create-r2fvl\" (UID: \"532ac93c-0c2c-4bca-b9f2-e4f066913dbf\") " pod="openstack/nova-cell1-db-create-r2fvl" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.684880 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glq8s\" (UniqueName: \"kubernetes.io/projected/532ac93c-0c2c-4bca-b9f2-e4f066913dbf-kube-api-access-glq8s\") pod \"nova-cell1-db-create-r2fvl\" (UID: \"532ac93c-0c2c-4bca-b9f2-e4f066913dbf\") " pod="openstack/nova-cell1-db-create-r2fvl" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.715612 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glq8s\" (UniqueName: \"kubernetes.io/projected/532ac93c-0c2c-4bca-b9f2-e4f066913dbf-kube-api-access-glq8s\") pod \"nova-cell1-db-create-r2fvl\" (UID: \"532ac93c-0c2c-4bca-b9f2-e4f066913dbf\") " pod="openstack/nova-cell1-db-create-r2fvl" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.755812 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-r2fvl" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.785947 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5d6f9192-b64d-4390-a4bd-4ff0a33bc31b-operator-scripts\") pod \"nova-cell0-56b5-account-create-update-98zzg\" (UID: \"5d6f9192-b64d-4390-a4bd-4ff0a33bc31b\") " pod="openstack/nova-cell0-56b5-account-create-update-98zzg" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.786629 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5d6f9192-b64d-4390-a4bd-4ff0a33bc31b-operator-scripts\") pod \"nova-cell0-56b5-account-create-update-98zzg\" (UID: \"5d6f9192-b64d-4390-a4bd-4ff0a33bc31b\") " pod="openstack/nova-cell0-56b5-account-create-update-98zzg" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.786689 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7fbe6119-47a9-4188-b8c0-a9e87d8ea09f-operator-scripts\") pod \"nova-cell1-db9a-account-create-update-kb927\" (UID: \"7fbe6119-47a9-4188-b8c0-a9e87d8ea09f\") " pod="openstack/nova-cell1-db9a-account-create-update-kb927" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.791365 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7fbe6119-47a9-4188-b8c0-a9e87d8ea09f-operator-scripts\") pod \"nova-cell1-db9a-account-create-update-kb927\" (UID: \"7fbe6119-47a9-4188-b8c0-a9e87d8ea09f\") " pod="openstack/nova-cell1-db9a-account-create-update-kb927" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.786768 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4p77\" (UniqueName: \"kubernetes.io/projected/5d6f9192-b64d-4390-a4bd-4ff0a33bc31b-kube-api-access-s4p77\") pod \"nova-cell0-56b5-account-create-update-98zzg\" (UID: \"5d6f9192-b64d-4390-a4bd-4ff0a33bc31b\") " pod="openstack/nova-cell0-56b5-account-create-update-98zzg" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.791548 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl84k\" (UniqueName: \"kubernetes.io/projected/7fbe6119-47a9-4188-b8c0-a9e87d8ea09f-kube-api-access-tl84k\") pod \"nova-cell1-db9a-account-create-update-kb927\" (UID: \"7fbe6119-47a9-4188-b8c0-a9e87d8ea09f\") " pod="openstack/nova-cell1-db9a-account-create-update-kb927" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.819883 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4p77\" (UniqueName: \"kubernetes.io/projected/5d6f9192-b64d-4390-a4bd-4ff0a33bc31b-kube-api-access-s4p77\") pod \"nova-cell0-56b5-account-create-update-98zzg\" (UID: \"5d6f9192-b64d-4390-a4bd-4ff0a33bc31b\") " pod="openstack/nova-cell0-56b5-account-create-update-98zzg" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.851509 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl84k\" (UniqueName: \"kubernetes.io/projected/7fbe6119-47a9-4188-b8c0-a9e87d8ea09f-kube-api-access-tl84k\") pod \"nova-cell1-db9a-account-create-update-kb927\" (UID: \"7fbe6119-47a9-4188-b8c0-a9e87d8ea09f\") " pod="openstack/nova-cell1-db9a-account-create-update-kb927" Dec 06 16:04:25 crc kubenswrapper[4813]: I1206 16:04:25.854734 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-56b5-account-create-update-98zzg" Dec 06 16:04:26 crc kubenswrapper[4813]: I1206 16:04:26.066005 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-mxrr6"] Dec 06 16:04:26 crc kubenswrapper[4813]: I1206 16:04:26.112967 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db9a-account-create-update-kb927" Dec 06 16:04:26 crc kubenswrapper[4813]: I1206 16:04:26.231372 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-bbc668d58-h8w67" podUID="4bdc8329-162b-4592-b98e-ec7eda7f2ce1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Dec 06 16:04:26 crc kubenswrapper[4813]: I1206 16:04:26.322519 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-mxrr6" event={"ID":"a06cadf5-56f1-439f-8deb-2aaedba66c8a","Type":"ContainerStarted","Data":"28d6f6ab8daecf57a453142ea98d454a61308e45ab7058bb40f44a3c32905cf9"} Dec 06 16:04:26 crc kubenswrapper[4813]: I1206 16:04:26.338967 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-3fbf-account-create-update-ksdbr"] Dec 06 16:04:26 crc kubenswrapper[4813]: I1206 16:04:26.385839 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-t86b9"] Dec 06 16:04:26 crc kubenswrapper[4813]: I1206 16:04:26.470826 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-r2fvl"] Dec 06 16:04:26 crc kubenswrapper[4813]: I1206 16:04:26.701940 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-56b5-account-create-update-98zzg"] Dec 06 16:04:26 crc kubenswrapper[4813]: I1206 16:04:26.785120 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db9a-account-create-update-kb927"] Dec 06 16:04:26 crc kubenswrapper[4813]: W1206 16:04:26.785724 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7fbe6119_47a9_4188_b8c0_a9e87d8ea09f.slice/crio-50c53482e2e572939beb784749e146649e50dc84716c11db16f8eea7f0eba1b4 WatchSource:0}: Error finding container 50c53482e2e572939beb784749e146649e50dc84716c11db16f8eea7f0eba1b4: Status 404 returned error can't find the container with id 50c53482e2e572939beb784749e146649e50dc84716c11db16f8eea7f0eba1b4 Dec 06 16:04:27 crc kubenswrapper[4813]: I1206 16:04:27.331564 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-r2fvl" event={"ID":"532ac93c-0c2c-4bca-b9f2-e4f066913dbf","Type":"ContainerStarted","Data":"7fa069863f89a6a74b90af08db253cba8b1d2424f9d923060e7159e60040f086"} Dec 06 16:04:27 crc kubenswrapper[4813]: I1206 16:04:27.332953 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3fbf-account-create-update-ksdbr" event={"ID":"dc35c201-8f5a-418b-a615-c04dff409d8d","Type":"ContainerStarted","Data":"0f1286cc3f731a84e12d2f463b641e3c3a6da4c4b571f268463a4a6661a7818e"} Dec 06 16:04:27 crc kubenswrapper[4813]: I1206 16:04:27.334681 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-t86b9" event={"ID":"47a481f6-b873-4920-a273-6e65f8d6503a","Type":"ContainerStarted","Data":"50660e9f6af1e822329407bacebb4af6192e880eae3b454d9d0d520ec553aac9"} Dec 06 16:04:27 crc kubenswrapper[4813]: I1206 16:04:27.335919 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db9a-account-create-update-kb927" event={"ID":"7fbe6119-47a9-4188-b8c0-a9e87d8ea09f","Type":"ContainerStarted","Data":"50c53482e2e572939beb784749e146649e50dc84716c11db16f8eea7f0eba1b4"} Dec 06 16:04:29 crc kubenswrapper[4813]: I1206 16:04:29.232669 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 16:04:29 crc kubenswrapper[4813]: I1206 16:04:29.242071 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="5bccad5f-06ef-46af-b5a7-f4b761b02550" containerName="glance-httpd" containerID="cri-o://9af44f16657483a9f9bcb4d8995264621f9f361deb554a563ffaf3ecbe22f6d3" gracePeriod=30 Dec 06 16:04:29 crc kubenswrapper[4813]: I1206 16:04:29.242088 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="5bccad5f-06ef-46af-b5a7-f4b761b02550" containerName="glance-log" containerID="cri-o://f7750f8c7a2adbdd23c031fe8057997104f67f39c52c41c224f654e6a65985e1" gracePeriod=30 Dec 06 16:04:29 crc kubenswrapper[4813]: I1206 16:04:29.353097 4813 generic.go:334] "Generic (PLEG): container finished" podID="a06cadf5-56f1-439f-8deb-2aaedba66c8a" containerID="2b018776e90533759f52261401f35f1e5b333dd629be589509b2457f0a31dd8e" exitCode=0 Dec 06 16:04:29 crc kubenswrapper[4813]: I1206 16:04:29.353155 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-mxrr6" event={"ID":"a06cadf5-56f1-439f-8deb-2aaedba66c8a","Type":"ContainerDied","Data":"2b018776e90533759f52261401f35f1e5b333dd629be589509b2457f0a31dd8e"} Dec 06 16:04:29 crc kubenswrapper[4813]: I1206 16:04:29.355540 4813 generic.go:334] "Generic (PLEG): container finished" podID="532ac93c-0c2c-4bca-b9f2-e4f066913dbf" containerID="4cdc7c62d685ecd333aa00daac5609ba7f02b43330e726059f3482413810fd1e" exitCode=0 Dec 06 16:04:29 crc kubenswrapper[4813]: I1206 16:04:29.355580 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-r2fvl" event={"ID":"532ac93c-0c2c-4bca-b9f2-e4f066913dbf","Type":"ContainerDied","Data":"4cdc7c62d685ecd333aa00daac5609ba7f02b43330e726059f3482413810fd1e"} Dec 06 16:04:29 crc kubenswrapper[4813]: I1206 16:04:29.358730 4813 generic.go:334] "Generic (PLEG): container finished" podID="5d6f9192-b64d-4390-a4bd-4ff0a33bc31b" containerID="75bb59db87ccb27e3651904a042c755b9be081706b6842c2bd2ad04120ed6287" exitCode=0 Dec 06 16:04:29 crc kubenswrapper[4813]: I1206 16:04:29.358774 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-56b5-account-create-update-98zzg" event={"ID":"5d6f9192-b64d-4390-a4bd-4ff0a33bc31b","Type":"ContainerDied","Data":"75bb59db87ccb27e3651904a042c755b9be081706b6842c2bd2ad04120ed6287"} Dec 06 16:04:29 crc kubenswrapper[4813]: I1206 16:04:29.358789 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-56b5-account-create-update-98zzg" event={"ID":"5d6f9192-b64d-4390-a4bd-4ff0a33bc31b","Type":"ContainerStarted","Data":"75d659017f4cb91b69b3ef79e20c4c971d8306c9f84c1c645534bab0dffe3473"} Dec 06 16:04:29 crc kubenswrapper[4813]: I1206 16:04:29.360226 4813 generic.go:334] "Generic (PLEG): container finished" podID="dc35c201-8f5a-418b-a615-c04dff409d8d" containerID="adda738648174e8dcbda04a4c13cbb2f263e4a82465ea6f49b4c3082ed252d21" exitCode=0 Dec 06 16:04:29 crc kubenswrapper[4813]: I1206 16:04:29.360278 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3fbf-account-create-update-ksdbr" event={"ID":"dc35c201-8f5a-418b-a615-c04dff409d8d","Type":"ContainerDied","Data":"adda738648174e8dcbda04a4c13cbb2f263e4a82465ea6f49b4c3082ed252d21"} Dec 06 16:04:29 crc kubenswrapper[4813]: I1206 16:04:29.362702 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54c36b42-5bd6-4c47-92c2-80ac1be6646f","Type":"ContainerStarted","Data":"b20e53c94a1f94f50de016261aa60fdf5411325c88684b50fa30b59e5ddb85ec"} Dec 06 16:04:29 crc kubenswrapper[4813]: I1206 16:04:29.362713 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="54c36b42-5bd6-4c47-92c2-80ac1be6646f" containerName="ceilometer-central-agent" containerID="cri-o://cd8d1a683279343806f365da7844e6ec389a6e50ac444b8e9f03178e55a3f5da" gracePeriod=30 Dec 06 16:04:29 crc kubenswrapper[4813]: I1206 16:04:29.362841 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="54c36b42-5bd6-4c47-92c2-80ac1be6646f" containerName="sg-core" containerID="cri-o://5de8e29fd04a40c2e172e5c514b2c0058f683d1db695f38c26b1ff48816a5908" gracePeriod=30 Dec 06 16:04:29 crc kubenswrapper[4813]: I1206 16:04:29.362863 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="54c36b42-5bd6-4c47-92c2-80ac1be6646f" containerName="ceilometer-notification-agent" containerID="cri-o://f59deba9a17c1dd00efa199a4e897a0ab13ac64172156ed7d55d30bf2e48f493" gracePeriod=30 Dec 06 16:04:29 crc kubenswrapper[4813]: I1206 16:04:29.362890 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="54c36b42-5bd6-4c47-92c2-80ac1be6646f" containerName="proxy-httpd" containerID="cri-o://b20e53c94a1f94f50de016261aa60fdf5411325c88684b50fa30b59e5ddb85ec" gracePeriod=30 Dec 06 16:04:29 crc kubenswrapper[4813]: I1206 16:04:29.363030 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 16:04:29 crc kubenswrapper[4813]: I1206 16:04:29.365528 4813 generic.go:334] "Generic (PLEG): container finished" podID="47a481f6-b873-4920-a273-6e65f8d6503a" containerID="74d957793a0594fe543c66f2360a457565a7f5aef669690ca0c52d34fe724784" exitCode=0 Dec 06 16:04:29 crc kubenswrapper[4813]: I1206 16:04:29.365583 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-t86b9" event={"ID":"47a481f6-b873-4920-a273-6e65f8d6503a","Type":"ContainerDied","Data":"74d957793a0594fe543c66f2360a457565a7f5aef669690ca0c52d34fe724784"} Dec 06 16:04:29 crc kubenswrapper[4813]: I1206 16:04:29.371797 4813 generic.go:334] "Generic (PLEG): container finished" podID="7fbe6119-47a9-4188-b8c0-a9e87d8ea09f" containerID="c8f9099fcb752af009bd425b0fbf2b15bdbfb56e518110eec0f2d4b689301885" exitCode=0 Dec 06 16:04:29 crc kubenswrapper[4813]: I1206 16:04:29.371835 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db9a-account-create-update-kb927" event={"ID":"7fbe6119-47a9-4188-b8c0-a9e87d8ea09f","Type":"ContainerDied","Data":"c8f9099fcb752af009bd425b0fbf2b15bdbfb56e518110eec0f2d4b689301885"} Dec 06 16:04:29 crc kubenswrapper[4813]: I1206 16:04:29.420561 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.769102926 podStartE2EDuration="7.420546807s" podCreationTimestamp="2025-12-06 16:04:22 +0000 UTC" firstStartedPulling="2025-12-06 16:04:23.169505965 +0000 UTC m=+1103.060385541" lastFinishedPulling="2025-12-06 16:04:28.820949846 +0000 UTC m=+1108.711829422" observedRunningTime="2025-12-06 16:04:29.416764268 +0000 UTC m=+1109.307643844" watchObservedRunningTime="2025-12-06 16:04:29.420546807 +0000 UTC m=+1109.311426383" Dec 06 16:04:30 crc kubenswrapper[4813]: I1206 16:04:30.090362 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 16:04:30 crc kubenswrapper[4813]: I1206 16:04:30.090795 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="6aa78964-46a9-4784-8d7a-9762bac90670" containerName="glance-log" containerID="cri-o://aac4a1741df703d35aaee5da55da2f600ff92c65b049582b59acb6099c789212" gracePeriod=30 Dec 06 16:04:30 crc kubenswrapper[4813]: I1206 16:04:30.090916 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="6aa78964-46a9-4784-8d7a-9762bac90670" containerName="glance-httpd" containerID="cri-o://8d56b17b6ec7d2287eae2b2f27075a41e25619c58499a939e57419b069771a85" gracePeriod=30 Dec 06 16:04:30 crc kubenswrapper[4813]: I1206 16:04:30.387235 4813 generic.go:334] "Generic (PLEG): container finished" podID="6aa78964-46a9-4784-8d7a-9762bac90670" containerID="aac4a1741df703d35aaee5da55da2f600ff92c65b049582b59acb6099c789212" exitCode=143 Dec 06 16:04:30 crc kubenswrapper[4813]: I1206 16:04:30.387320 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6aa78964-46a9-4784-8d7a-9762bac90670","Type":"ContainerDied","Data":"aac4a1741df703d35aaee5da55da2f600ff92c65b049582b59acb6099c789212"} Dec 06 16:04:30 crc kubenswrapper[4813]: I1206 16:04:30.390126 4813 generic.go:334] "Generic (PLEG): container finished" podID="5bccad5f-06ef-46af-b5a7-f4b761b02550" containerID="f7750f8c7a2adbdd23c031fe8057997104f67f39c52c41c224f654e6a65985e1" exitCode=143 Dec 06 16:04:30 crc kubenswrapper[4813]: I1206 16:04:30.390179 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5bccad5f-06ef-46af-b5a7-f4b761b02550","Type":"ContainerDied","Data":"f7750f8c7a2adbdd23c031fe8057997104f67f39c52c41c224f654e6a65985e1"} Dec 06 16:04:30 crc kubenswrapper[4813]: I1206 16:04:30.392234 4813 generic.go:334] "Generic (PLEG): container finished" podID="54c36b42-5bd6-4c47-92c2-80ac1be6646f" containerID="b20e53c94a1f94f50de016261aa60fdf5411325c88684b50fa30b59e5ddb85ec" exitCode=0 Dec 06 16:04:30 crc kubenswrapper[4813]: I1206 16:04:30.392274 4813 generic.go:334] "Generic (PLEG): container finished" podID="54c36b42-5bd6-4c47-92c2-80ac1be6646f" containerID="5de8e29fd04a40c2e172e5c514b2c0058f683d1db695f38c26b1ff48816a5908" exitCode=2 Dec 06 16:04:30 crc kubenswrapper[4813]: I1206 16:04:30.392283 4813 generic.go:334] "Generic (PLEG): container finished" podID="54c36b42-5bd6-4c47-92c2-80ac1be6646f" containerID="f59deba9a17c1dd00efa199a4e897a0ab13ac64172156ed7d55d30bf2e48f493" exitCode=0 Dec 06 16:04:30 crc kubenswrapper[4813]: I1206 16:04:30.392280 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54c36b42-5bd6-4c47-92c2-80ac1be6646f","Type":"ContainerDied","Data":"b20e53c94a1f94f50de016261aa60fdf5411325c88684b50fa30b59e5ddb85ec"} Dec 06 16:04:30 crc kubenswrapper[4813]: I1206 16:04:30.392413 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54c36b42-5bd6-4c47-92c2-80ac1be6646f","Type":"ContainerDied","Data":"5de8e29fd04a40c2e172e5c514b2c0058f683d1db695f38c26b1ff48816a5908"} Dec 06 16:04:30 crc kubenswrapper[4813]: I1206 16:04:30.392439 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54c36b42-5bd6-4c47-92c2-80ac1be6646f","Type":"ContainerDied","Data":"f59deba9a17c1dd00efa199a4e897a0ab13ac64172156ed7d55d30bf2e48f493"} Dec 06 16:04:30 crc kubenswrapper[4813]: I1206 16:04:30.706581 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-t86b9" Dec 06 16:04:30 crc kubenswrapper[4813]: I1206 16:04:30.849482 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdh5n\" (UniqueName: \"kubernetes.io/projected/47a481f6-b873-4920-a273-6e65f8d6503a-kube-api-access-vdh5n\") pod \"47a481f6-b873-4920-a273-6e65f8d6503a\" (UID: \"47a481f6-b873-4920-a273-6e65f8d6503a\") " Dec 06 16:04:30 crc kubenswrapper[4813]: I1206 16:04:30.849807 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47a481f6-b873-4920-a273-6e65f8d6503a-operator-scripts\") pod \"47a481f6-b873-4920-a273-6e65f8d6503a\" (UID: \"47a481f6-b873-4920-a273-6e65f8d6503a\") " Dec 06 16:04:30 crc kubenswrapper[4813]: I1206 16:04:30.850304 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47a481f6-b873-4920-a273-6e65f8d6503a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "47a481f6-b873-4920-a273-6e65f8d6503a" (UID: "47a481f6-b873-4920-a273-6e65f8d6503a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:04:30 crc kubenswrapper[4813]: I1206 16:04:30.850710 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47a481f6-b873-4920-a273-6e65f8d6503a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:30 crc kubenswrapper[4813]: I1206 16:04:30.857287 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47a481f6-b873-4920-a273-6e65f8d6503a-kube-api-access-vdh5n" (OuterVolumeSpecName: "kube-api-access-vdh5n") pod "47a481f6-b873-4920-a273-6e65f8d6503a" (UID: "47a481f6-b873-4920-a273-6e65f8d6503a"). InnerVolumeSpecName "kube-api-access-vdh5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:04:30 crc kubenswrapper[4813]: I1206 16:04:30.944207 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3fbf-account-create-update-ksdbr" Dec 06 16:04:30 crc kubenswrapper[4813]: I1206 16:04:30.952097 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdh5n\" (UniqueName: \"kubernetes.io/projected/47a481f6-b873-4920-a273-6e65f8d6503a-kube-api-access-vdh5n\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.058824 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghjhf\" (UniqueName: \"kubernetes.io/projected/dc35c201-8f5a-418b-a615-c04dff409d8d-kube-api-access-ghjhf\") pod \"dc35c201-8f5a-418b-a615-c04dff409d8d\" (UID: \"dc35c201-8f5a-418b-a615-c04dff409d8d\") " Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.058953 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc35c201-8f5a-418b-a615-c04dff409d8d-operator-scripts\") pod \"dc35c201-8f5a-418b-a615-c04dff409d8d\" (UID: \"dc35c201-8f5a-418b-a615-c04dff409d8d\") " Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.059826 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc35c201-8f5a-418b-a615-c04dff409d8d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dc35c201-8f5a-418b-a615-c04dff409d8d" (UID: "dc35c201-8f5a-418b-a615-c04dff409d8d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.064965 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc35c201-8f5a-418b-a615-c04dff409d8d-kube-api-access-ghjhf" (OuterVolumeSpecName: "kube-api-access-ghjhf") pod "dc35c201-8f5a-418b-a615-c04dff409d8d" (UID: "dc35c201-8f5a-418b-a615-c04dff409d8d"). InnerVolumeSpecName "kube-api-access-ghjhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.157441 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-r2fvl" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.161567 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghjhf\" (UniqueName: \"kubernetes.io/projected/dc35c201-8f5a-418b-a615-c04dff409d8d-kube-api-access-ghjhf\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.161596 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc35c201-8f5a-418b-a615-c04dff409d8d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.169237 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db9a-account-create-update-kb927" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.192167 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-mxrr6" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.215817 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-56b5-account-create-update-98zzg" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.267570 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tl84k\" (UniqueName: \"kubernetes.io/projected/7fbe6119-47a9-4188-b8c0-a9e87d8ea09f-kube-api-access-tl84k\") pod \"7fbe6119-47a9-4188-b8c0-a9e87d8ea09f\" (UID: \"7fbe6119-47a9-4188-b8c0-a9e87d8ea09f\") " Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.267605 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/532ac93c-0c2c-4bca-b9f2-e4f066913dbf-operator-scripts\") pod \"532ac93c-0c2c-4bca-b9f2-e4f066913dbf\" (UID: \"532ac93c-0c2c-4bca-b9f2-e4f066913dbf\") " Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.267653 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glq8s\" (UniqueName: \"kubernetes.io/projected/532ac93c-0c2c-4bca-b9f2-e4f066913dbf-kube-api-access-glq8s\") pod \"532ac93c-0c2c-4bca-b9f2-e4f066913dbf\" (UID: \"532ac93c-0c2c-4bca-b9f2-e4f066913dbf\") " Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.267823 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7fbe6119-47a9-4188-b8c0-a9e87d8ea09f-operator-scripts\") pod \"7fbe6119-47a9-4188-b8c0-a9e87d8ea09f\" (UID: \"7fbe6119-47a9-4188-b8c0-a9e87d8ea09f\") " Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.268487 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fbe6119-47a9-4188-b8c0-a9e87d8ea09f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7fbe6119-47a9-4188-b8c0-a9e87d8ea09f" (UID: "7fbe6119-47a9-4188-b8c0-a9e87d8ea09f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.269351 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/532ac93c-0c2c-4bca-b9f2-e4f066913dbf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "532ac93c-0c2c-4bca-b9f2-e4f066913dbf" (UID: "532ac93c-0c2c-4bca-b9f2-e4f066913dbf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.279002 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fbe6119-47a9-4188-b8c0-a9e87d8ea09f-kube-api-access-tl84k" (OuterVolumeSpecName: "kube-api-access-tl84k") pod "7fbe6119-47a9-4188-b8c0-a9e87d8ea09f" (UID: "7fbe6119-47a9-4188-b8c0-a9e87d8ea09f"). InnerVolumeSpecName "kube-api-access-tl84k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.279485 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/532ac93c-0c2c-4bca-b9f2-e4f066913dbf-kube-api-access-glq8s" (OuterVolumeSpecName: "kube-api-access-glq8s") pod "532ac93c-0c2c-4bca-b9f2-e4f066913dbf" (UID: "532ac93c-0c2c-4bca-b9f2-e4f066913dbf"). InnerVolumeSpecName "kube-api-access-glq8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.369890 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a06cadf5-56f1-439f-8deb-2aaedba66c8a-operator-scripts\") pod \"a06cadf5-56f1-439f-8deb-2aaedba66c8a\" (UID: \"a06cadf5-56f1-439f-8deb-2aaedba66c8a\") " Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.370008 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qsrtv\" (UniqueName: \"kubernetes.io/projected/a06cadf5-56f1-439f-8deb-2aaedba66c8a-kube-api-access-qsrtv\") pod \"a06cadf5-56f1-439f-8deb-2aaedba66c8a\" (UID: \"a06cadf5-56f1-439f-8deb-2aaedba66c8a\") " Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.370055 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4p77\" (UniqueName: \"kubernetes.io/projected/5d6f9192-b64d-4390-a4bd-4ff0a33bc31b-kube-api-access-s4p77\") pod \"5d6f9192-b64d-4390-a4bd-4ff0a33bc31b\" (UID: \"5d6f9192-b64d-4390-a4bd-4ff0a33bc31b\") " Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.370076 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5d6f9192-b64d-4390-a4bd-4ff0a33bc31b-operator-scripts\") pod \"5d6f9192-b64d-4390-a4bd-4ff0a33bc31b\" (UID: \"5d6f9192-b64d-4390-a4bd-4ff0a33bc31b\") " Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.371817 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7fbe6119-47a9-4188-b8c0-a9e87d8ea09f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.371913 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tl84k\" (UniqueName: \"kubernetes.io/projected/7fbe6119-47a9-4188-b8c0-a9e87d8ea09f-kube-api-access-tl84k\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.371926 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/532ac93c-0c2c-4bca-b9f2-e4f066913dbf-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.371935 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glq8s\" (UniqueName: \"kubernetes.io/projected/532ac93c-0c2c-4bca-b9f2-e4f066913dbf-kube-api-access-glq8s\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.372596 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d6f9192-b64d-4390-a4bd-4ff0a33bc31b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5d6f9192-b64d-4390-a4bd-4ff0a33bc31b" (UID: "5d6f9192-b64d-4390-a4bd-4ff0a33bc31b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.374146 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a06cadf5-56f1-439f-8deb-2aaedba66c8a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a06cadf5-56f1-439f-8deb-2aaedba66c8a" (UID: "a06cadf5-56f1-439f-8deb-2aaedba66c8a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.374200 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a06cadf5-56f1-439f-8deb-2aaedba66c8a-kube-api-access-qsrtv" (OuterVolumeSpecName: "kube-api-access-qsrtv") pod "a06cadf5-56f1-439f-8deb-2aaedba66c8a" (UID: "a06cadf5-56f1-439f-8deb-2aaedba66c8a"). InnerVolumeSpecName "kube-api-access-qsrtv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.374584 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d6f9192-b64d-4390-a4bd-4ff0a33bc31b-kube-api-access-s4p77" (OuterVolumeSpecName: "kube-api-access-s4p77") pod "5d6f9192-b64d-4390-a4bd-4ff0a33bc31b" (UID: "5d6f9192-b64d-4390-a4bd-4ff0a33bc31b"). InnerVolumeSpecName "kube-api-access-s4p77". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.414869 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-r2fvl" event={"ID":"532ac93c-0c2c-4bca-b9f2-e4f066913dbf","Type":"ContainerDied","Data":"7fa069863f89a6a74b90af08db253cba8b1d2424f9d923060e7159e60040f086"} Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.414930 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7fa069863f89a6a74b90af08db253cba8b1d2424f9d923060e7159e60040f086" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.415008 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-r2fvl" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.430883 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-56b5-account-create-update-98zzg" event={"ID":"5d6f9192-b64d-4390-a4bd-4ff0a33bc31b","Type":"ContainerDied","Data":"75d659017f4cb91b69b3ef79e20c4c971d8306c9f84c1c645534bab0dffe3473"} Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.430920 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75d659017f4cb91b69b3ef79e20c4c971d8306c9f84c1c645534bab0dffe3473" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.430975 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-56b5-account-create-update-98zzg" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.436412 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3fbf-account-create-update-ksdbr" event={"ID":"dc35c201-8f5a-418b-a615-c04dff409d8d","Type":"ContainerDied","Data":"0f1286cc3f731a84e12d2f463b641e3c3a6da4c4b571f268463a4a6661a7818e"} Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.436586 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f1286cc3f731a84e12d2f463b641e3c3a6da4c4b571f268463a4a6661a7818e" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.436623 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3fbf-account-create-update-ksdbr" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.445356 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-bbc668d58-h8w67" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.457615 4813 generic.go:334] "Generic (PLEG): container finished" podID="4bdc8329-162b-4592-b98e-ec7eda7f2ce1" containerID="859333a385a1f24c723ad6a5689eae7998a7664c905e033301fe7108cfa531de" exitCode=137 Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.457686 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-bbc668d58-h8w67" event={"ID":"4bdc8329-162b-4592-b98e-ec7eda7f2ce1","Type":"ContainerDied","Data":"859333a385a1f24c723ad6a5689eae7998a7664c905e033301fe7108cfa531de"} Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.457720 4813 scope.go:117] "RemoveContainer" containerID="3307007b53464dbc4d2a32eb29a1577325f2d8df6457bc6adf713e100058d865" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.461549 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-t86b9" event={"ID":"47a481f6-b873-4920-a273-6e65f8d6503a","Type":"ContainerDied","Data":"50660e9f6af1e822329407bacebb4af6192e880eae3b454d9d0d520ec553aac9"} Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.461577 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50660e9f6af1e822329407bacebb4af6192e880eae3b454d9d0d520ec553aac9" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.461640 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-t86b9" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.482986 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qsrtv\" (UniqueName: \"kubernetes.io/projected/a06cadf5-56f1-439f-8deb-2aaedba66c8a-kube-api-access-qsrtv\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.483030 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4p77\" (UniqueName: \"kubernetes.io/projected/5d6f9192-b64d-4390-a4bd-4ff0a33bc31b-kube-api-access-s4p77\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.483043 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5d6f9192-b64d-4390-a4bd-4ff0a33bc31b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.483053 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a06cadf5-56f1-439f-8deb-2aaedba66c8a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.487833 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db9a-account-create-update-kb927" event={"ID":"7fbe6119-47a9-4188-b8c0-a9e87d8ea09f","Type":"ContainerDied","Data":"50c53482e2e572939beb784749e146649e50dc84716c11db16f8eea7f0eba1b4"} Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.487860 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50c53482e2e572939beb784749e146649e50dc84716c11db16f8eea7f0eba1b4" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.487912 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db9a-account-create-update-kb927" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.497065 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-mxrr6" event={"ID":"a06cadf5-56f1-439f-8deb-2aaedba66c8a","Type":"ContainerDied","Data":"28d6f6ab8daecf57a453142ea98d454a61308e45ab7058bb40f44a3c32905cf9"} Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.497104 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28d6f6ab8daecf57a453142ea98d454a61308e45ab7058bb40f44a3c32905cf9" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.497159 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-mxrr6" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.585762 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-horizon-secret-key\") pod \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\" (UID: \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\") " Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.585849 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-combined-ca-bundle\") pod \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\" (UID: \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\") " Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.585893 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-config-data\") pod \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\" (UID: \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\") " Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.585918 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-horizon-tls-certs\") pod \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\" (UID: \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\") " Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.585989 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-logs\") pod \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\" (UID: \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\") " Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.586050 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ttbz\" (UniqueName: \"kubernetes.io/projected/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-kube-api-access-8ttbz\") pod \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\" (UID: \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\") " Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.586087 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-scripts\") pod \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\" (UID: \"4bdc8329-162b-4592-b98e-ec7eda7f2ce1\") " Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.588966 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-logs" (OuterVolumeSpecName: "logs") pod "4bdc8329-162b-4592-b98e-ec7eda7f2ce1" (UID: "4bdc8329-162b-4592-b98e-ec7eda7f2ce1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.593890 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-kube-api-access-8ttbz" (OuterVolumeSpecName: "kube-api-access-8ttbz") pod "4bdc8329-162b-4592-b98e-ec7eda7f2ce1" (UID: "4bdc8329-162b-4592-b98e-ec7eda7f2ce1"). InnerVolumeSpecName "kube-api-access-8ttbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.597593 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "4bdc8329-162b-4592-b98e-ec7eda7f2ce1" (UID: "4bdc8329-162b-4592-b98e-ec7eda7f2ce1"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.638267 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-config-data" (OuterVolumeSpecName: "config-data") pod "4bdc8329-162b-4592-b98e-ec7eda7f2ce1" (UID: "4bdc8329-162b-4592-b98e-ec7eda7f2ce1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.639674 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-scripts" (OuterVolumeSpecName: "scripts") pod "4bdc8329-162b-4592-b98e-ec7eda7f2ce1" (UID: "4bdc8329-162b-4592-b98e-ec7eda7f2ce1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.641154 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4bdc8329-162b-4592-b98e-ec7eda7f2ce1" (UID: "4bdc8329-162b-4592-b98e-ec7eda7f2ce1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.692774 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.692818 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.692829 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-logs\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.692837 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ttbz\" (UniqueName: \"kubernetes.io/projected/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-kube-api-access-8ttbz\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.692847 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.692857 4813 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.693022 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "4bdc8329-162b-4592-b98e-ec7eda7f2ce1" (UID: "4bdc8329-162b-4592-b98e-ec7eda7f2ce1"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.750947 4813 scope.go:117] "RemoveContainer" containerID="859333a385a1f24c723ad6a5689eae7998a7664c905e033301fe7108cfa531de" Dec 06 16:04:31 crc kubenswrapper[4813]: I1206 16:04:31.794156 4813 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bdc8329-162b-4592-b98e-ec7eda7f2ce1-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:32 crc kubenswrapper[4813]: I1206 16:04:32.524919 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-bbc668d58-h8w67" event={"ID":"4bdc8329-162b-4592-b98e-ec7eda7f2ce1","Type":"ContainerDied","Data":"eb41bb3b1c841fb4eb3841ac4832b4adeefc0eb6a531069ccd6879a1ba4e6ef9"} Dec 06 16:04:32 crc kubenswrapper[4813]: I1206 16:04:32.525018 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-bbc668d58-h8w67" Dec 06 16:04:32 crc kubenswrapper[4813]: I1206 16:04:32.534800 4813 generic.go:334] "Generic (PLEG): container finished" podID="5bccad5f-06ef-46af-b5a7-f4b761b02550" containerID="9af44f16657483a9f9bcb4d8995264621f9f361deb554a563ffaf3ecbe22f6d3" exitCode=0 Dec 06 16:04:32 crc kubenswrapper[4813]: I1206 16:04:32.534835 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5bccad5f-06ef-46af-b5a7-f4b761b02550","Type":"ContainerDied","Data":"9af44f16657483a9f9bcb4d8995264621f9f361deb554a563ffaf3ecbe22f6d3"} Dec 06 16:04:32 crc kubenswrapper[4813]: I1206 16:04:32.569638 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-bbc668d58-h8w67"] Dec 06 16:04:32 crc kubenswrapper[4813]: I1206 16:04:32.595487 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-bbc668d58-h8w67"] Dec 06 16:04:32 crc kubenswrapper[4813]: I1206 16:04:32.875993 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 16:04:32 crc kubenswrapper[4813]: I1206 16:04:32.935998 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bccad5f-06ef-46af-b5a7-f4b761b02550-logs\") pod \"5bccad5f-06ef-46af-b5a7-f4b761b02550\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " Dec 06 16:04:32 crc kubenswrapper[4813]: I1206 16:04:32.936080 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lplm\" (UniqueName: \"kubernetes.io/projected/5bccad5f-06ef-46af-b5a7-f4b761b02550-kube-api-access-7lplm\") pod \"5bccad5f-06ef-46af-b5a7-f4b761b02550\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " Dec 06 16:04:32 crc kubenswrapper[4813]: I1206 16:04:32.936132 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bccad5f-06ef-46af-b5a7-f4b761b02550-config-data\") pod \"5bccad5f-06ef-46af-b5a7-f4b761b02550\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " Dec 06 16:04:32 crc kubenswrapper[4813]: I1206 16:04:32.936241 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bccad5f-06ef-46af-b5a7-f4b761b02550-combined-ca-bundle\") pod \"5bccad5f-06ef-46af-b5a7-f4b761b02550\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " Dec 06 16:04:32 crc kubenswrapper[4813]: I1206 16:04:32.936304 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5bccad5f-06ef-46af-b5a7-f4b761b02550-httpd-run\") pod \"5bccad5f-06ef-46af-b5a7-f4b761b02550\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " Dec 06 16:04:32 crc kubenswrapper[4813]: I1206 16:04:32.936321 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bccad5f-06ef-46af-b5a7-f4b761b02550-public-tls-certs\") pod \"5bccad5f-06ef-46af-b5a7-f4b761b02550\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " Dec 06 16:04:32 crc kubenswrapper[4813]: I1206 16:04:32.936356 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bccad5f-06ef-46af-b5a7-f4b761b02550-scripts\") pod \"5bccad5f-06ef-46af-b5a7-f4b761b02550\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " Dec 06 16:04:32 crc kubenswrapper[4813]: I1206 16:04:32.936415 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"5bccad5f-06ef-46af-b5a7-f4b761b02550\" (UID: \"5bccad5f-06ef-46af-b5a7-f4b761b02550\") " Dec 06 16:04:32 crc kubenswrapper[4813]: I1206 16:04:32.937005 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bccad5f-06ef-46af-b5a7-f4b761b02550-logs" (OuterVolumeSpecName: "logs") pod "5bccad5f-06ef-46af-b5a7-f4b761b02550" (UID: "5bccad5f-06ef-46af-b5a7-f4b761b02550"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:04:32 crc kubenswrapper[4813]: I1206 16:04:32.939675 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bccad5f-06ef-46af-b5a7-f4b761b02550-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5bccad5f-06ef-46af-b5a7-f4b761b02550" (UID: "5bccad5f-06ef-46af-b5a7-f4b761b02550"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:04:32 crc kubenswrapper[4813]: I1206 16:04:32.944251 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "5bccad5f-06ef-46af-b5a7-f4b761b02550" (UID: "5bccad5f-06ef-46af-b5a7-f4b761b02550"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 16:04:32 crc kubenswrapper[4813]: I1206 16:04:32.951453 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bccad5f-06ef-46af-b5a7-f4b761b02550-scripts" (OuterVolumeSpecName: "scripts") pod "5bccad5f-06ef-46af-b5a7-f4b761b02550" (UID: "5bccad5f-06ef-46af-b5a7-f4b761b02550"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:04:32 crc kubenswrapper[4813]: I1206 16:04:32.960449 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bccad5f-06ef-46af-b5a7-f4b761b02550-kube-api-access-7lplm" (OuterVolumeSpecName: "kube-api-access-7lplm") pod "5bccad5f-06ef-46af-b5a7-f4b761b02550" (UID: "5bccad5f-06ef-46af-b5a7-f4b761b02550"). InnerVolumeSpecName "kube-api-access-7lplm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:04:32 crc kubenswrapper[4813]: I1206 16:04:32.994384 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bccad5f-06ef-46af-b5a7-f4b761b02550-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5bccad5f-06ef-46af-b5a7-f4b761b02550" (UID: "5bccad5f-06ef-46af-b5a7-f4b761b02550"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.043444 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bccad5f-06ef-46af-b5a7-f4b761b02550-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.043471 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5bccad5f-06ef-46af-b5a7-f4b761b02550-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.043480 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bccad5f-06ef-46af-b5a7-f4b761b02550-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.043499 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.043508 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bccad5f-06ef-46af-b5a7-f4b761b02550-logs\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.043516 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lplm\" (UniqueName: \"kubernetes.io/projected/5bccad5f-06ef-46af-b5a7-f4b761b02550-kube-api-access-7lplm\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.044419 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bccad5f-06ef-46af-b5a7-f4b761b02550-config-data" (OuterVolumeSpecName: "config-data") pod "5bccad5f-06ef-46af-b5a7-f4b761b02550" (UID: "5bccad5f-06ef-46af-b5a7-f4b761b02550"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.061996 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.089049 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bccad5f-06ef-46af-b5a7-f4b761b02550-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5bccad5f-06ef-46af-b5a7-f4b761b02550" (UID: "5bccad5f-06ef-46af-b5a7-f4b761b02550"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.145346 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bccad5f-06ef-46af-b5a7-f4b761b02550-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.145378 4813 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bccad5f-06ef-46af-b5a7-f4b761b02550-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.145390 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.545823 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5bccad5f-06ef-46af-b5a7-f4b761b02550","Type":"ContainerDied","Data":"29c2ce8c03fa93fedc67caadfcfaa11b0a06cb8e004ff0dd414d77465c54aa80"} Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.545853 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.546074 4813 scope.go:117] "RemoveContainer" containerID="9af44f16657483a9f9bcb4d8995264621f9f361deb554a563ffaf3ecbe22f6d3" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.548773 4813 generic.go:334] "Generic (PLEG): container finished" podID="6aa78964-46a9-4784-8d7a-9762bac90670" containerID="8d56b17b6ec7d2287eae2b2f27075a41e25619c58499a939e57419b069771a85" exitCode=0 Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.548802 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6aa78964-46a9-4784-8d7a-9762bac90670","Type":"ContainerDied","Data":"8d56b17b6ec7d2287eae2b2f27075a41e25619c58499a939e57419b069771a85"} Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.589940 4813 scope.go:117] "RemoveContainer" containerID="f7750f8c7a2adbdd23c031fe8057997104f67f39c52c41c224f654e6a65985e1" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.610385 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.637508 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.663339 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 16:04:33 crc kubenswrapper[4813]: E1206 16:04:33.663740 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bccad5f-06ef-46af-b5a7-f4b761b02550" containerName="glance-httpd" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.663752 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bccad5f-06ef-46af-b5a7-f4b761b02550" containerName="glance-httpd" Dec 06 16:04:33 crc kubenswrapper[4813]: E1206 16:04:33.663773 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a06cadf5-56f1-439f-8deb-2aaedba66c8a" containerName="mariadb-database-create" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.663780 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a06cadf5-56f1-439f-8deb-2aaedba66c8a" containerName="mariadb-database-create" Dec 06 16:04:33 crc kubenswrapper[4813]: E1206 16:04:33.663790 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bdc8329-162b-4592-b98e-ec7eda7f2ce1" containerName="horizon" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.663796 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bdc8329-162b-4592-b98e-ec7eda7f2ce1" containerName="horizon" Dec 06 16:04:33 crc kubenswrapper[4813]: E1206 16:04:33.663813 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bccad5f-06ef-46af-b5a7-f4b761b02550" containerName="glance-log" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.663819 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bccad5f-06ef-46af-b5a7-f4b761b02550" containerName="glance-log" Dec 06 16:04:33 crc kubenswrapper[4813]: E1206 16:04:33.663832 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d6f9192-b64d-4390-a4bd-4ff0a33bc31b" containerName="mariadb-account-create-update" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.663838 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d6f9192-b64d-4390-a4bd-4ff0a33bc31b" containerName="mariadb-account-create-update" Dec 06 16:04:33 crc kubenswrapper[4813]: E1206 16:04:33.663850 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47a481f6-b873-4920-a273-6e65f8d6503a" containerName="mariadb-database-create" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.663855 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="47a481f6-b873-4920-a273-6e65f8d6503a" containerName="mariadb-database-create" Dec 06 16:04:33 crc kubenswrapper[4813]: E1206 16:04:33.663863 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bdc8329-162b-4592-b98e-ec7eda7f2ce1" containerName="horizon-log" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.663868 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bdc8329-162b-4592-b98e-ec7eda7f2ce1" containerName="horizon-log" Dec 06 16:04:33 crc kubenswrapper[4813]: E1206 16:04:33.663877 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="532ac93c-0c2c-4bca-b9f2-e4f066913dbf" containerName="mariadb-database-create" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.663883 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="532ac93c-0c2c-4bca-b9f2-e4f066913dbf" containerName="mariadb-database-create" Dec 06 16:04:33 crc kubenswrapper[4813]: E1206 16:04:33.663894 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc35c201-8f5a-418b-a615-c04dff409d8d" containerName="mariadb-account-create-update" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.663900 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc35c201-8f5a-418b-a615-c04dff409d8d" containerName="mariadb-account-create-update" Dec 06 16:04:33 crc kubenswrapper[4813]: E1206 16:04:33.663914 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fbe6119-47a9-4188-b8c0-a9e87d8ea09f" containerName="mariadb-account-create-update" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.663919 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fbe6119-47a9-4188-b8c0-a9e87d8ea09f" containerName="mariadb-account-create-update" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.664111 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bdc8329-162b-4592-b98e-ec7eda7f2ce1" containerName="horizon" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.664124 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bccad5f-06ef-46af-b5a7-f4b761b02550" containerName="glance-httpd" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.664138 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d6f9192-b64d-4390-a4bd-4ff0a33bc31b" containerName="mariadb-account-create-update" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.664149 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="47a481f6-b873-4920-a273-6e65f8d6503a" containerName="mariadb-database-create" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.664161 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a06cadf5-56f1-439f-8deb-2aaedba66c8a" containerName="mariadb-database-create" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.664170 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bccad5f-06ef-46af-b5a7-f4b761b02550" containerName="glance-log" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.664192 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="532ac93c-0c2c-4bca-b9f2-e4f066913dbf" containerName="mariadb-database-create" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.664202 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fbe6119-47a9-4188-b8c0-a9e87d8ea09f" containerName="mariadb-account-create-update" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.664213 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc35c201-8f5a-418b-a615-c04dff409d8d" containerName="mariadb-account-create-update" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.664227 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bdc8329-162b-4592-b98e-ec7eda7f2ce1" containerName="horizon-log" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.665112 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.669070 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.670636 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.674021 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.722986 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.754706 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-js27q\" (UniqueName: \"kubernetes.io/projected/6aa78964-46a9-4784-8d7a-9762bac90670-kube-api-access-js27q\") pod \"6aa78964-46a9-4784-8d7a-9762bac90670\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.755199 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6aa78964-46a9-4784-8d7a-9762bac90670-httpd-run\") pod \"6aa78964-46a9-4784-8d7a-9762bac90670\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.755323 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6aa78964-46a9-4784-8d7a-9762bac90670-config-data\") pod \"6aa78964-46a9-4784-8d7a-9762bac90670\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.755421 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6aa78964-46a9-4784-8d7a-9762bac90670-scripts\") pod \"6aa78964-46a9-4784-8d7a-9762bac90670\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.755502 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aa78964-46a9-4784-8d7a-9762bac90670-combined-ca-bundle\") pod \"6aa78964-46a9-4784-8d7a-9762bac90670\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.755585 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6aa78964-46a9-4784-8d7a-9762bac90670-logs\") pod \"6aa78964-46a9-4784-8d7a-9762bac90670\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.755685 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6aa78964-46a9-4784-8d7a-9762bac90670-internal-tls-certs\") pod \"6aa78964-46a9-4784-8d7a-9762bac90670\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.755756 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"6aa78964-46a9-4784-8d7a-9762bac90670\" (UID: \"6aa78964-46a9-4784-8d7a-9762bac90670\") " Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.755935 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee3e963f-6982-4cd8-87d2-4eceaec10be3-scripts\") pod \"glance-default-external-api-0\" (UID: \"ee3e963f-6982-4cd8-87d2-4eceaec10be3\") " pod="openstack/glance-default-external-api-0" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.756008 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee3e963f-6982-4cd8-87d2-4eceaec10be3-config-data\") pod \"glance-default-external-api-0\" (UID: \"ee3e963f-6982-4cd8-87d2-4eceaec10be3\") " pod="openstack/glance-default-external-api-0" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.756361 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee3e963f-6982-4cd8-87d2-4eceaec10be3-logs\") pod \"glance-default-external-api-0\" (UID: \"ee3e963f-6982-4cd8-87d2-4eceaec10be3\") " pod="openstack/glance-default-external-api-0" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.756473 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ee3e963f-6982-4cd8-87d2-4eceaec10be3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ee3e963f-6982-4cd8-87d2-4eceaec10be3\") " pod="openstack/glance-default-external-api-0" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.756541 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"ee3e963f-6982-4cd8-87d2-4eceaec10be3\") " pod="openstack/glance-default-external-api-0" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.756647 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l87s\" (UniqueName: \"kubernetes.io/projected/ee3e963f-6982-4cd8-87d2-4eceaec10be3-kube-api-access-6l87s\") pod \"glance-default-external-api-0\" (UID: \"ee3e963f-6982-4cd8-87d2-4eceaec10be3\") " pod="openstack/glance-default-external-api-0" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.756750 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee3e963f-6982-4cd8-87d2-4eceaec10be3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ee3e963f-6982-4cd8-87d2-4eceaec10be3\") " pod="openstack/glance-default-external-api-0" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.756838 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee3e963f-6982-4cd8-87d2-4eceaec10be3-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ee3e963f-6982-4cd8-87d2-4eceaec10be3\") " pod="openstack/glance-default-external-api-0" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.763578 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6aa78964-46a9-4784-8d7a-9762bac90670-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "6aa78964-46a9-4784-8d7a-9762bac90670" (UID: "6aa78964-46a9-4784-8d7a-9762bac90670"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.764388 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6aa78964-46a9-4784-8d7a-9762bac90670-kube-api-access-js27q" (OuterVolumeSpecName: "kube-api-access-js27q") pod "6aa78964-46a9-4784-8d7a-9762bac90670" (UID: "6aa78964-46a9-4784-8d7a-9762bac90670"). InnerVolumeSpecName "kube-api-access-js27q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.764593 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6aa78964-46a9-4784-8d7a-9762bac90670-logs" (OuterVolumeSpecName: "logs") pod "6aa78964-46a9-4784-8d7a-9762bac90670" (UID: "6aa78964-46a9-4784-8d7a-9762bac90670"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.768702 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "6aa78964-46a9-4784-8d7a-9762bac90670" (UID: "6aa78964-46a9-4784-8d7a-9762bac90670"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.769212 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6aa78964-46a9-4784-8d7a-9762bac90670-scripts" (OuterVolumeSpecName: "scripts") pod "6aa78964-46a9-4784-8d7a-9762bac90670" (UID: "6aa78964-46a9-4784-8d7a-9762bac90670"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.840389 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6aa78964-46a9-4784-8d7a-9762bac90670-config-data" (OuterVolumeSpecName: "config-data") pod "6aa78964-46a9-4784-8d7a-9762bac90670" (UID: "6aa78964-46a9-4784-8d7a-9762bac90670"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.852328 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6aa78964-46a9-4784-8d7a-9762bac90670-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6aa78964-46a9-4784-8d7a-9762bac90670" (UID: "6aa78964-46a9-4784-8d7a-9762bac90670"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.859180 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ee3e963f-6982-4cd8-87d2-4eceaec10be3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ee3e963f-6982-4cd8-87d2-4eceaec10be3\") " pod="openstack/glance-default-external-api-0" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.859213 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"ee3e963f-6982-4cd8-87d2-4eceaec10be3\") " pod="openstack/glance-default-external-api-0" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.859357 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l87s\" (UniqueName: \"kubernetes.io/projected/ee3e963f-6982-4cd8-87d2-4eceaec10be3-kube-api-access-6l87s\") pod \"glance-default-external-api-0\" (UID: \"ee3e963f-6982-4cd8-87d2-4eceaec10be3\") " pod="openstack/glance-default-external-api-0" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.859405 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee3e963f-6982-4cd8-87d2-4eceaec10be3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ee3e963f-6982-4cd8-87d2-4eceaec10be3\") " pod="openstack/glance-default-external-api-0" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.859435 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee3e963f-6982-4cd8-87d2-4eceaec10be3-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ee3e963f-6982-4cd8-87d2-4eceaec10be3\") " pod="openstack/glance-default-external-api-0" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.859473 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee3e963f-6982-4cd8-87d2-4eceaec10be3-scripts\") pod \"glance-default-external-api-0\" (UID: \"ee3e963f-6982-4cd8-87d2-4eceaec10be3\") " pod="openstack/glance-default-external-api-0" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.859491 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee3e963f-6982-4cd8-87d2-4eceaec10be3-config-data\") pod \"glance-default-external-api-0\" (UID: \"ee3e963f-6982-4cd8-87d2-4eceaec10be3\") " pod="openstack/glance-default-external-api-0" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.859510 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee3e963f-6982-4cd8-87d2-4eceaec10be3-logs\") pod \"glance-default-external-api-0\" (UID: \"ee3e963f-6982-4cd8-87d2-4eceaec10be3\") " pod="openstack/glance-default-external-api-0" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.859556 4813 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6aa78964-46a9-4784-8d7a-9762bac90670-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.859576 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.859585 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-js27q\" (UniqueName: \"kubernetes.io/projected/6aa78964-46a9-4784-8d7a-9762bac90670-kube-api-access-js27q\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.859594 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6aa78964-46a9-4784-8d7a-9762bac90670-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.859604 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6aa78964-46a9-4784-8d7a-9762bac90670-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.859614 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6aa78964-46a9-4784-8d7a-9762bac90670-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.859622 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6aa78964-46a9-4784-8d7a-9762bac90670-logs\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.866454 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee3e963f-6982-4cd8-87d2-4eceaec10be3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ee3e963f-6982-4cd8-87d2-4eceaec10be3\") " pod="openstack/glance-default-external-api-0" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.866711 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ee3e963f-6982-4cd8-87d2-4eceaec10be3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ee3e963f-6982-4cd8-87d2-4eceaec10be3\") " pod="openstack/glance-default-external-api-0" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.866871 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"ee3e963f-6982-4cd8-87d2-4eceaec10be3\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-external-api-0" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.868553 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee3e963f-6982-4cd8-87d2-4eceaec10be3-logs\") pod \"glance-default-external-api-0\" (UID: \"ee3e963f-6982-4cd8-87d2-4eceaec10be3\") " pod="openstack/glance-default-external-api-0" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.880841 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee3e963f-6982-4cd8-87d2-4eceaec10be3-scripts\") pod \"glance-default-external-api-0\" (UID: \"ee3e963f-6982-4cd8-87d2-4eceaec10be3\") " pod="openstack/glance-default-external-api-0" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.883700 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee3e963f-6982-4cd8-87d2-4eceaec10be3-config-data\") pod \"glance-default-external-api-0\" (UID: \"ee3e963f-6982-4cd8-87d2-4eceaec10be3\") " pod="openstack/glance-default-external-api-0" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.885823 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee3e963f-6982-4cd8-87d2-4eceaec10be3-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ee3e963f-6982-4cd8-87d2-4eceaec10be3\") " pod="openstack/glance-default-external-api-0" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.890630 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l87s\" (UniqueName: \"kubernetes.io/projected/ee3e963f-6982-4cd8-87d2-4eceaec10be3-kube-api-access-6l87s\") pod \"glance-default-external-api-0\" (UID: \"ee3e963f-6982-4cd8-87d2-4eceaec10be3\") " pod="openstack/glance-default-external-api-0" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.906441 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.918763 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6aa78964-46a9-4784-8d7a-9762bac90670-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6aa78964-46a9-4784-8d7a-9762bac90670" (UID: "6aa78964-46a9-4784-8d7a-9762bac90670"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.936870 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"ee3e963f-6982-4cd8-87d2-4eceaec10be3\") " pod="openstack/glance-default-external-api-0" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.962442 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aa78964-46a9-4784-8d7a-9762bac90670-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.962678 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:33 crc kubenswrapper[4813]: I1206 16:04:33.989077 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.495664 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bdc8329-162b-4592-b98e-ec7eda7f2ce1" path="/var/lib/kubelet/pods/4bdc8329-162b-4592-b98e-ec7eda7f2ce1/volumes" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.496579 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bccad5f-06ef-46af-b5a7-f4b761b02550" path="/var/lib/kubelet/pods/5bccad5f-06ef-46af-b5a7-f4b761b02550/volumes" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.568829 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6aa78964-46a9-4784-8d7a-9762bac90670","Type":"ContainerDied","Data":"29adfd7844b15dbad3e74bde5271f288e44cf33a80d9e3c51e32eb734510cad5"} Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.569658 4813 scope.go:117] "RemoveContainer" containerID="8d56b17b6ec7d2287eae2b2f27075a41e25619c58499a939e57419b069771a85" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.569366 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.618526 4813 scope.go:117] "RemoveContainer" containerID="aac4a1741df703d35aaee5da55da2f600ff92c65b049582b59acb6099c789212" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.621916 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.638333 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.648505 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.658319 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 16:04:34 crc kubenswrapper[4813]: E1206 16:04:34.662434 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6aa78964-46a9-4784-8d7a-9762bac90670" containerName="glance-httpd" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.662454 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="6aa78964-46a9-4784-8d7a-9762bac90670" containerName="glance-httpd" Dec 06 16:04:34 crc kubenswrapper[4813]: E1206 16:04:34.662472 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6aa78964-46a9-4784-8d7a-9762bac90670" containerName="glance-log" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.662478 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="6aa78964-46a9-4784-8d7a-9762bac90670" containerName="glance-log" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.662655 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="6aa78964-46a9-4784-8d7a-9762bac90670" containerName="glance-log" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.662671 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="6aa78964-46a9-4784-8d7a-9762bac90670" containerName="glance-httpd" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.665029 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.667648 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.672636 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.672774 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.678123 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7crs\" (UniqueName: \"kubernetes.io/projected/ec17345c-3d8e-4981-945c-173fb39aab99-kube-api-access-s7crs\") pod \"glance-default-internal-api-0\" (UID: \"ec17345c-3d8e-4981-945c-173fb39aab99\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.678210 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec17345c-3d8e-4981-945c-173fb39aab99-logs\") pod \"glance-default-internal-api-0\" (UID: \"ec17345c-3d8e-4981-945c-173fb39aab99\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.678229 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec17345c-3d8e-4981-945c-173fb39aab99-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ec17345c-3d8e-4981-945c-173fb39aab99\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.678314 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ec17345c-3d8e-4981-945c-173fb39aab99-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ec17345c-3d8e-4981-945c-173fb39aab99\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.678343 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec17345c-3d8e-4981-945c-173fb39aab99-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ec17345c-3d8e-4981-945c-173fb39aab99\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.678367 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"ec17345c-3d8e-4981-945c-173fb39aab99\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.678395 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec17345c-3d8e-4981-945c-173fb39aab99-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ec17345c-3d8e-4981-945c-173fb39aab99\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.678410 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec17345c-3d8e-4981-945c-173fb39aab99-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ec17345c-3d8e-4981-945c-173fb39aab99\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.780789 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"ec17345c-3d8e-4981-945c-173fb39aab99\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.780874 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec17345c-3d8e-4981-945c-173fb39aab99-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ec17345c-3d8e-4981-945c-173fb39aab99\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.780895 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec17345c-3d8e-4981-945c-173fb39aab99-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ec17345c-3d8e-4981-945c-173fb39aab99\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.780935 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7crs\" (UniqueName: \"kubernetes.io/projected/ec17345c-3d8e-4981-945c-173fb39aab99-kube-api-access-s7crs\") pod \"glance-default-internal-api-0\" (UID: \"ec17345c-3d8e-4981-945c-173fb39aab99\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.781040 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec17345c-3d8e-4981-945c-173fb39aab99-logs\") pod \"glance-default-internal-api-0\" (UID: \"ec17345c-3d8e-4981-945c-173fb39aab99\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.781056 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec17345c-3d8e-4981-945c-173fb39aab99-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ec17345c-3d8e-4981-945c-173fb39aab99\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.781091 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ec17345c-3d8e-4981-945c-173fb39aab99-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ec17345c-3d8e-4981-945c-173fb39aab99\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.781134 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec17345c-3d8e-4981-945c-173fb39aab99-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ec17345c-3d8e-4981-945c-173fb39aab99\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.781662 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"ec17345c-3d8e-4981-945c-173fb39aab99\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.784739 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec17345c-3d8e-4981-945c-173fb39aab99-logs\") pod \"glance-default-internal-api-0\" (UID: \"ec17345c-3d8e-4981-945c-173fb39aab99\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.784786 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ec17345c-3d8e-4981-945c-173fb39aab99-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ec17345c-3d8e-4981-945c-173fb39aab99\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.795986 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec17345c-3d8e-4981-945c-173fb39aab99-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ec17345c-3d8e-4981-945c-173fb39aab99\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.796748 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec17345c-3d8e-4981-945c-173fb39aab99-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ec17345c-3d8e-4981-945c-173fb39aab99\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.798654 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec17345c-3d8e-4981-945c-173fb39aab99-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ec17345c-3d8e-4981-945c-173fb39aab99\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.801440 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7crs\" (UniqueName: \"kubernetes.io/projected/ec17345c-3d8e-4981-945c-173fb39aab99-kube-api-access-s7crs\") pod \"glance-default-internal-api-0\" (UID: \"ec17345c-3d8e-4981-945c-173fb39aab99\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.804209 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec17345c-3d8e-4981-945c-173fb39aab99-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ec17345c-3d8e-4981-945c-173fb39aab99\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:04:34 crc kubenswrapper[4813]: I1206 16:04:34.831018 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"ec17345c-3d8e-4981-945c-173fb39aab99\") " pod="openstack/glance-default-internal-api-0" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.001066 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.250334 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.403218 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/54c36b42-5bd6-4c47-92c2-80ac1be6646f-sg-core-conf-yaml\") pod \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\" (UID: \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\") " Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.403320 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54c36b42-5bd6-4c47-92c2-80ac1be6646f-run-httpd\") pod \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\" (UID: \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\") " Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.403371 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54c36b42-5bd6-4c47-92c2-80ac1be6646f-combined-ca-bundle\") pod \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\" (UID: \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\") " Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.403402 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54c36b42-5bd6-4c47-92c2-80ac1be6646f-config-data\") pod \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\" (UID: \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\") " Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.403474 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54c36b42-5bd6-4c47-92c2-80ac1be6646f-log-httpd\") pod \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\" (UID: \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\") " Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.403535 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54c36b42-5bd6-4c47-92c2-80ac1be6646f-scripts\") pod \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\" (UID: \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\") " Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.403594 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9sjw5\" (UniqueName: \"kubernetes.io/projected/54c36b42-5bd6-4c47-92c2-80ac1be6646f-kube-api-access-9sjw5\") pod \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\" (UID: \"54c36b42-5bd6-4c47-92c2-80ac1be6646f\") " Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.404357 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54c36b42-5bd6-4c47-92c2-80ac1be6646f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "54c36b42-5bd6-4c47-92c2-80ac1be6646f" (UID: "54c36b42-5bd6-4c47-92c2-80ac1be6646f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.404633 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54c36b42-5bd6-4c47-92c2-80ac1be6646f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "54c36b42-5bd6-4c47-92c2-80ac1be6646f" (UID: "54c36b42-5bd6-4c47-92c2-80ac1be6646f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.421743 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54c36b42-5bd6-4c47-92c2-80ac1be6646f-kube-api-access-9sjw5" (OuterVolumeSpecName: "kube-api-access-9sjw5") pod "54c36b42-5bd6-4c47-92c2-80ac1be6646f" (UID: "54c36b42-5bd6-4c47-92c2-80ac1be6646f"). InnerVolumeSpecName "kube-api-access-9sjw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.425335 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54c36b42-5bd6-4c47-92c2-80ac1be6646f-scripts" (OuterVolumeSpecName: "scripts") pod "54c36b42-5bd6-4c47-92c2-80ac1be6646f" (UID: "54c36b42-5bd6-4c47-92c2-80ac1be6646f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.496610 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54c36b42-5bd6-4c47-92c2-80ac1be6646f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "54c36b42-5bd6-4c47-92c2-80ac1be6646f" (UID: "54c36b42-5bd6-4c47-92c2-80ac1be6646f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.506248 4813 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54c36b42-5bd6-4c47-92c2-80ac1be6646f-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.506293 4813 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54c36b42-5bd6-4c47-92c2-80ac1be6646f-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.506304 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54c36b42-5bd6-4c47-92c2-80ac1be6646f-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.506313 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9sjw5\" (UniqueName: \"kubernetes.io/projected/54c36b42-5bd6-4c47-92c2-80ac1be6646f-kube-api-access-9sjw5\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.506322 4813 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/54c36b42-5bd6-4c47-92c2-80ac1be6646f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.534090 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54c36b42-5bd6-4c47-92c2-80ac1be6646f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "54c36b42-5bd6-4c47-92c2-80ac1be6646f" (UID: "54c36b42-5bd6-4c47-92c2-80ac1be6646f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.584479 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54c36b42-5bd6-4c47-92c2-80ac1be6646f-config-data" (OuterVolumeSpecName: "config-data") pod "54c36b42-5bd6-4c47-92c2-80ac1be6646f" (UID: "54c36b42-5bd6-4c47-92c2-80ac1be6646f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.607092 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54c36b42-5bd6-4c47-92c2-80ac1be6646f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.607123 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54c36b42-5bd6-4c47-92c2-80ac1be6646f-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.607742 4813 generic.go:334] "Generic (PLEG): container finished" podID="54c36b42-5bd6-4c47-92c2-80ac1be6646f" containerID="cd8d1a683279343806f365da7844e6ec389a6e50ac444b8e9f03178e55a3f5da" exitCode=0 Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.607788 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54c36b42-5bd6-4c47-92c2-80ac1be6646f","Type":"ContainerDied","Data":"cd8d1a683279343806f365da7844e6ec389a6e50ac444b8e9f03178e55a3f5da"} Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.607814 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54c36b42-5bd6-4c47-92c2-80ac1be6646f","Type":"ContainerDied","Data":"1ad6ab4359e25a54c11e3c95cf56e03b248766e5faac36d0faad3c482fef4ee3"} Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.607831 4813 scope.go:117] "RemoveContainer" containerID="b20e53c94a1f94f50de016261aa60fdf5411325c88684b50fa30b59e5ddb85ec" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.607918 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.616583 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ee3e963f-6982-4cd8-87d2-4eceaec10be3","Type":"ContainerStarted","Data":"6ff5156522e4af06e9a9126af952a416af46dada6fea3fd4645475ae0308de29"} Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.616620 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ee3e963f-6982-4cd8-87d2-4eceaec10be3","Type":"ContainerStarted","Data":"dd54fb8418b9a335d38588e7f6ac30f81d655401f817f9d1bfdf759bb2b3e597"} Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.641601 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.645500 4813 scope.go:117] "RemoveContainer" containerID="5de8e29fd04a40c2e172e5c514b2c0058f683d1db695f38c26b1ff48816a5908" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.652448 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.678179 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:04:35 crc kubenswrapper[4813]: E1206 16:04:35.684860 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54c36b42-5bd6-4c47-92c2-80ac1be6646f" containerName="sg-core" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.684895 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="54c36b42-5bd6-4c47-92c2-80ac1be6646f" containerName="sg-core" Dec 06 16:04:35 crc kubenswrapper[4813]: E1206 16:04:35.684908 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54c36b42-5bd6-4c47-92c2-80ac1be6646f" containerName="ceilometer-notification-agent" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.684915 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="54c36b42-5bd6-4c47-92c2-80ac1be6646f" containerName="ceilometer-notification-agent" Dec 06 16:04:35 crc kubenswrapper[4813]: E1206 16:04:35.684924 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54c36b42-5bd6-4c47-92c2-80ac1be6646f" containerName="ceilometer-central-agent" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.684930 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="54c36b42-5bd6-4c47-92c2-80ac1be6646f" containerName="ceilometer-central-agent" Dec 06 16:04:35 crc kubenswrapper[4813]: E1206 16:04:35.684970 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54c36b42-5bd6-4c47-92c2-80ac1be6646f" containerName="proxy-httpd" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.684977 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="54c36b42-5bd6-4c47-92c2-80ac1be6646f" containerName="proxy-httpd" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.685134 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="54c36b42-5bd6-4c47-92c2-80ac1be6646f" containerName="sg-core" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.685149 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="54c36b42-5bd6-4c47-92c2-80ac1be6646f" containerName="ceilometer-central-agent" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.685162 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="54c36b42-5bd6-4c47-92c2-80ac1be6646f" containerName="proxy-httpd" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.685174 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="54c36b42-5bd6-4c47-92c2-80ac1be6646f" containerName="ceilometer-notification-agent" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.686747 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.694124 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.694434 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.695286 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.695697 4813 scope.go:117] "RemoveContainer" containerID="f59deba9a17c1dd00efa199a4e897a0ab13ac64172156ed7d55d30bf2e48f493" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.796992 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.810722 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhqjp\" (UniqueName: \"kubernetes.io/projected/68cf1c73-04c1-4776-a91d-2189db97e9ed-kube-api-access-rhqjp\") pod \"ceilometer-0\" (UID: \"68cf1c73-04c1-4776-a91d-2189db97e9ed\") " pod="openstack/ceilometer-0" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.810787 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68cf1c73-04c1-4776-a91d-2189db97e9ed-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"68cf1c73-04c1-4776-a91d-2189db97e9ed\") " pod="openstack/ceilometer-0" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.810808 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/68cf1c73-04c1-4776-a91d-2189db97e9ed-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"68cf1c73-04c1-4776-a91d-2189db97e9ed\") " pod="openstack/ceilometer-0" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.810867 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68cf1c73-04c1-4776-a91d-2189db97e9ed-run-httpd\") pod \"ceilometer-0\" (UID: \"68cf1c73-04c1-4776-a91d-2189db97e9ed\") " pod="openstack/ceilometer-0" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.810899 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68cf1c73-04c1-4776-a91d-2189db97e9ed-scripts\") pod \"ceilometer-0\" (UID: \"68cf1c73-04c1-4776-a91d-2189db97e9ed\") " pod="openstack/ceilometer-0" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.810940 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68cf1c73-04c1-4776-a91d-2189db97e9ed-log-httpd\") pod \"ceilometer-0\" (UID: \"68cf1c73-04c1-4776-a91d-2189db97e9ed\") " pod="openstack/ceilometer-0" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.810964 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68cf1c73-04c1-4776-a91d-2189db97e9ed-config-data\") pod \"ceilometer-0\" (UID: \"68cf1c73-04c1-4776-a91d-2189db97e9ed\") " pod="openstack/ceilometer-0" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.830486 4813 scope.go:117] "RemoveContainer" containerID="cd8d1a683279343806f365da7844e6ec389a6e50ac444b8e9f03178e55a3f5da" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.852333 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-pn5s8"] Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.853462 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-pn5s8" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.868898 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-pn5s8"] Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.870746 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.871014 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.871143 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-9sldw" Dec 06 16:04:35 crc kubenswrapper[4813]: W1206 16:04:35.904667 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec17345c_3d8e_4981_945c_173fb39aab99.slice/crio-06878f6dc6c87274750edb6df195b46ddbcee4075be3c54a6b1a04119aa97803 WatchSource:0}: Error finding container 06878f6dc6c87274750edb6df195b46ddbcee4075be3c54a6b1a04119aa97803: Status 404 returned error can't find the container with id 06878f6dc6c87274750edb6df195b46ddbcee4075be3c54a6b1a04119aa97803 Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.912149 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68cf1c73-04c1-4776-a91d-2189db97e9ed-config-data\") pod \"ceilometer-0\" (UID: \"68cf1c73-04c1-4776-a91d-2189db97e9ed\") " pod="openstack/ceilometer-0" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.912192 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhqjp\" (UniqueName: \"kubernetes.io/projected/68cf1c73-04c1-4776-a91d-2189db97e9ed-kube-api-access-rhqjp\") pod \"ceilometer-0\" (UID: \"68cf1c73-04c1-4776-a91d-2189db97e9ed\") " pod="openstack/ceilometer-0" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.912243 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68cf1c73-04c1-4776-a91d-2189db97e9ed-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"68cf1c73-04c1-4776-a91d-2189db97e9ed\") " pod="openstack/ceilometer-0" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.912274 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/68cf1c73-04c1-4776-a91d-2189db97e9ed-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"68cf1c73-04c1-4776-a91d-2189db97e9ed\") " pod="openstack/ceilometer-0" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.912335 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68cf1c73-04c1-4776-a91d-2189db97e9ed-run-httpd\") pod \"ceilometer-0\" (UID: \"68cf1c73-04c1-4776-a91d-2189db97e9ed\") " pod="openstack/ceilometer-0" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.912360 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68cf1c73-04c1-4776-a91d-2189db97e9ed-scripts\") pod \"ceilometer-0\" (UID: \"68cf1c73-04c1-4776-a91d-2189db97e9ed\") " pod="openstack/ceilometer-0" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.912399 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68cf1c73-04c1-4776-a91d-2189db97e9ed-log-httpd\") pod \"ceilometer-0\" (UID: \"68cf1c73-04c1-4776-a91d-2189db97e9ed\") " pod="openstack/ceilometer-0" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.912846 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68cf1c73-04c1-4776-a91d-2189db97e9ed-log-httpd\") pod \"ceilometer-0\" (UID: \"68cf1c73-04c1-4776-a91d-2189db97e9ed\") " pod="openstack/ceilometer-0" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.912868 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68cf1c73-04c1-4776-a91d-2189db97e9ed-run-httpd\") pod \"ceilometer-0\" (UID: \"68cf1c73-04c1-4776-a91d-2189db97e9ed\") " pod="openstack/ceilometer-0" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.914376 4813 scope.go:117] "RemoveContainer" containerID="b20e53c94a1f94f50de016261aa60fdf5411325c88684b50fa30b59e5ddb85ec" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.919084 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/68cf1c73-04c1-4776-a91d-2189db97e9ed-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"68cf1c73-04c1-4776-a91d-2189db97e9ed\") " pod="openstack/ceilometer-0" Dec 06 16:04:35 crc kubenswrapper[4813]: E1206 16:04:35.922694 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b20e53c94a1f94f50de016261aa60fdf5411325c88684b50fa30b59e5ddb85ec\": container with ID starting with b20e53c94a1f94f50de016261aa60fdf5411325c88684b50fa30b59e5ddb85ec not found: ID does not exist" containerID="b20e53c94a1f94f50de016261aa60fdf5411325c88684b50fa30b59e5ddb85ec" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.922758 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b20e53c94a1f94f50de016261aa60fdf5411325c88684b50fa30b59e5ddb85ec"} err="failed to get container status \"b20e53c94a1f94f50de016261aa60fdf5411325c88684b50fa30b59e5ddb85ec\": rpc error: code = NotFound desc = could not find container \"b20e53c94a1f94f50de016261aa60fdf5411325c88684b50fa30b59e5ddb85ec\": container with ID starting with b20e53c94a1f94f50de016261aa60fdf5411325c88684b50fa30b59e5ddb85ec not found: ID does not exist" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.922785 4813 scope.go:117] "RemoveContainer" containerID="5de8e29fd04a40c2e172e5c514b2c0058f683d1db695f38c26b1ff48816a5908" Dec 06 16:04:35 crc kubenswrapper[4813]: E1206 16:04:35.923271 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5de8e29fd04a40c2e172e5c514b2c0058f683d1db695f38c26b1ff48816a5908\": container with ID starting with 5de8e29fd04a40c2e172e5c514b2c0058f683d1db695f38c26b1ff48816a5908 not found: ID does not exist" containerID="5de8e29fd04a40c2e172e5c514b2c0058f683d1db695f38c26b1ff48816a5908" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.923317 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5de8e29fd04a40c2e172e5c514b2c0058f683d1db695f38c26b1ff48816a5908"} err="failed to get container status \"5de8e29fd04a40c2e172e5c514b2c0058f683d1db695f38c26b1ff48816a5908\": rpc error: code = NotFound desc = could not find container \"5de8e29fd04a40c2e172e5c514b2c0058f683d1db695f38c26b1ff48816a5908\": container with ID starting with 5de8e29fd04a40c2e172e5c514b2c0058f683d1db695f38c26b1ff48816a5908 not found: ID does not exist" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.923357 4813 scope.go:117] "RemoveContainer" containerID="f59deba9a17c1dd00efa199a4e897a0ab13ac64172156ed7d55d30bf2e48f493" Dec 06 16:04:35 crc kubenswrapper[4813]: E1206 16:04:35.924140 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f59deba9a17c1dd00efa199a4e897a0ab13ac64172156ed7d55d30bf2e48f493\": container with ID starting with f59deba9a17c1dd00efa199a4e897a0ab13ac64172156ed7d55d30bf2e48f493 not found: ID does not exist" containerID="f59deba9a17c1dd00efa199a4e897a0ab13ac64172156ed7d55d30bf2e48f493" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.924162 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f59deba9a17c1dd00efa199a4e897a0ab13ac64172156ed7d55d30bf2e48f493"} err="failed to get container status \"f59deba9a17c1dd00efa199a4e897a0ab13ac64172156ed7d55d30bf2e48f493\": rpc error: code = NotFound desc = could not find container \"f59deba9a17c1dd00efa199a4e897a0ab13ac64172156ed7d55d30bf2e48f493\": container with ID starting with f59deba9a17c1dd00efa199a4e897a0ab13ac64172156ed7d55d30bf2e48f493 not found: ID does not exist" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.924175 4813 scope.go:117] "RemoveContainer" containerID="cd8d1a683279343806f365da7844e6ec389a6e50ac444b8e9f03178e55a3f5da" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.924548 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68cf1c73-04c1-4776-a91d-2189db97e9ed-config-data\") pod \"ceilometer-0\" (UID: \"68cf1c73-04c1-4776-a91d-2189db97e9ed\") " pod="openstack/ceilometer-0" Dec 06 16:04:35 crc kubenswrapper[4813]: E1206 16:04:35.924627 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd8d1a683279343806f365da7844e6ec389a6e50ac444b8e9f03178e55a3f5da\": container with ID starting with cd8d1a683279343806f365da7844e6ec389a6e50ac444b8e9f03178e55a3f5da not found: ID does not exist" containerID="cd8d1a683279343806f365da7844e6ec389a6e50ac444b8e9f03178e55a3f5da" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.925175 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd8d1a683279343806f365da7844e6ec389a6e50ac444b8e9f03178e55a3f5da"} err="failed to get container status \"cd8d1a683279343806f365da7844e6ec389a6e50ac444b8e9f03178e55a3f5da\": rpc error: code = NotFound desc = could not find container \"cd8d1a683279343806f365da7844e6ec389a6e50ac444b8e9f03178e55a3f5da\": container with ID starting with cd8d1a683279343806f365da7844e6ec389a6e50ac444b8e9f03178e55a3f5da not found: ID does not exist" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.933454 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68cf1c73-04c1-4776-a91d-2189db97e9ed-scripts\") pod \"ceilometer-0\" (UID: \"68cf1c73-04c1-4776-a91d-2189db97e9ed\") " pod="openstack/ceilometer-0" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.937083 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhqjp\" (UniqueName: \"kubernetes.io/projected/68cf1c73-04c1-4776-a91d-2189db97e9ed-kube-api-access-rhqjp\") pod \"ceilometer-0\" (UID: \"68cf1c73-04c1-4776-a91d-2189db97e9ed\") " pod="openstack/ceilometer-0" Dec 06 16:04:35 crc kubenswrapper[4813]: I1206 16:04:35.949147 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68cf1c73-04c1-4776-a91d-2189db97e9ed-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"68cf1c73-04c1-4776-a91d-2189db97e9ed\") " pod="openstack/ceilometer-0" Dec 06 16:04:36 crc kubenswrapper[4813]: I1206 16:04:36.013608 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ff658cc-71bb-4df5-8ba8-1c0979e03d10-scripts\") pod \"nova-cell0-conductor-db-sync-pn5s8\" (UID: \"2ff658cc-71bb-4df5-8ba8-1c0979e03d10\") " pod="openstack/nova-cell0-conductor-db-sync-pn5s8" Dec 06 16:04:36 crc kubenswrapper[4813]: I1206 16:04:36.013687 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ff658cc-71bb-4df5-8ba8-1c0979e03d10-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-pn5s8\" (UID: \"2ff658cc-71bb-4df5-8ba8-1c0979e03d10\") " pod="openstack/nova-cell0-conductor-db-sync-pn5s8" Dec 06 16:04:36 crc kubenswrapper[4813]: I1206 16:04:36.013717 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ff658cc-71bb-4df5-8ba8-1c0979e03d10-config-data\") pod \"nova-cell0-conductor-db-sync-pn5s8\" (UID: \"2ff658cc-71bb-4df5-8ba8-1c0979e03d10\") " pod="openstack/nova-cell0-conductor-db-sync-pn5s8" Dec 06 16:04:36 crc kubenswrapper[4813]: I1206 16:04:36.014330 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbmlq\" (UniqueName: \"kubernetes.io/projected/2ff658cc-71bb-4df5-8ba8-1c0979e03d10-kube-api-access-gbmlq\") pod \"nova-cell0-conductor-db-sync-pn5s8\" (UID: \"2ff658cc-71bb-4df5-8ba8-1c0979e03d10\") " pod="openstack/nova-cell0-conductor-db-sync-pn5s8" Dec 06 16:04:36 crc kubenswrapper[4813]: I1206 16:04:36.024491 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 16:04:36 crc kubenswrapper[4813]: I1206 16:04:36.116110 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ff658cc-71bb-4df5-8ba8-1c0979e03d10-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-pn5s8\" (UID: \"2ff658cc-71bb-4df5-8ba8-1c0979e03d10\") " pod="openstack/nova-cell0-conductor-db-sync-pn5s8" Dec 06 16:04:36 crc kubenswrapper[4813]: I1206 16:04:36.116156 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ff658cc-71bb-4df5-8ba8-1c0979e03d10-config-data\") pod \"nova-cell0-conductor-db-sync-pn5s8\" (UID: \"2ff658cc-71bb-4df5-8ba8-1c0979e03d10\") " pod="openstack/nova-cell0-conductor-db-sync-pn5s8" Dec 06 16:04:36 crc kubenswrapper[4813]: I1206 16:04:36.116230 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbmlq\" (UniqueName: \"kubernetes.io/projected/2ff658cc-71bb-4df5-8ba8-1c0979e03d10-kube-api-access-gbmlq\") pod \"nova-cell0-conductor-db-sync-pn5s8\" (UID: \"2ff658cc-71bb-4df5-8ba8-1c0979e03d10\") " pod="openstack/nova-cell0-conductor-db-sync-pn5s8" Dec 06 16:04:36 crc kubenswrapper[4813]: I1206 16:04:36.116306 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ff658cc-71bb-4df5-8ba8-1c0979e03d10-scripts\") pod \"nova-cell0-conductor-db-sync-pn5s8\" (UID: \"2ff658cc-71bb-4df5-8ba8-1c0979e03d10\") " pod="openstack/nova-cell0-conductor-db-sync-pn5s8" Dec 06 16:04:36 crc kubenswrapper[4813]: I1206 16:04:36.120059 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ff658cc-71bb-4df5-8ba8-1c0979e03d10-scripts\") pod \"nova-cell0-conductor-db-sync-pn5s8\" (UID: \"2ff658cc-71bb-4df5-8ba8-1c0979e03d10\") " pod="openstack/nova-cell0-conductor-db-sync-pn5s8" Dec 06 16:04:36 crc kubenswrapper[4813]: I1206 16:04:36.121813 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ff658cc-71bb-4df5-8ba8-1c0979e03d10-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-pn5s8\" (UID: \"2ff658cc-71bb-4df5-8ba8-1c0979e03d10\") " pod="openstack/nova-cell0-conductor-db-sync-pn5s8" Dec 06 16:04:36 crc kubenswrapper[4813]: I1206 16:04:36.121911 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ff658cc-71bb-4df5-8ba8-1c0979e03d10-config-data\") pod \"nova-cell0-conductor-db-sync-pn5s8\" (UID: \"2ff658cc-71bb-4df5-8ba8-1c0979e03d10\") " pod="openstack/nova-cell0-conductor-db-sync-pn5s8" Dec 06 16:04:36 crc kubenswrapper[4813]: I1206 16:04:36.132388 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbmlq\" (UniqueName: \"kubernetes.io/projected/2ff658cc-71bb-4df5-8ba8-1c0979e03d10-kube-api-access-gbmlq\") pod \"nova-cell0-conductor-db-sync-pn5s8\" (UID: \"2ff658cc-71bb-4df5-8ba8-1c0979e03d10\") " pod="openstack/nova-cell0-conductor-db-sync-pn5s8" Dec 06 16:04:36 crc kubenswrapper[4813]: I1206 16:04:36.189880 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-pn5s8" Dec 06 16:04:36 crc kubenswrapper[4813]: I1206 16:04:36.498557 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54c36b42-5bd6-4c47-92c2-80ac1be6646f" path="/var/lib/kubelet/pods/54c36b42-5bd6-4c47-92c2-80ac1be6646f/volumes" Dec 06 16:04:36 crc kubenswrapper[4813]: I1206 16:04:36.499733 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6aa78964-46a9-4784-8d7a-9762bac90670" path="/var/lib/kubelet/pods/6aa78964-46a9-4784-8d7a-9762bac90670/volumes" Dec 06 16:04:36 crc kubenswrapper[4813]: W1206 16:04:36.511405 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68cf1c73_04c1_4776_a91d_2189db97e9ed.slice/crio-6da18f0ee2d1d6e141d273b01e75a7bf57a69862051391d87d8adfeb1061ba54 WatchSource:0}: Error finding container 6da18f0ee2d1d6e141d273b01e75a7bf57a69862051391d87d8adfeb1061ba54: Status 404 returned error can't find the container with id 6da18f0ee2d1d6e141d273b01e75a7bf57a69862051391d87d8adfeb1061ba54 Dec 06 16:04:36 crc kubenswrapper[4813]: I1206 16:04:36.512372 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:04:36 crc kubenswrapper[4813]: I1206 16:04:36.632425 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ee3e963f-6982-4cd8-87d2-4eceaec10be3","Type":"ContainerStarted","Data":"d3f17caa0c5e7593139ef8bfb61570b873b2baaf693fbf406f29c5fed7df1479"} Dec 06 16:04:36 crc kubenswrapper[4813]: I1206 16:04:36.634601 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ec17345c-3d8e-4981-945c-173fb39aab99","Type":"ContainerStarted","Data":"06878f6dc6c87274750edb6df195b46ddbcee4075be3c54a6b1a04119aa97803"} Dec 06 16:04:36 crc kubenswrapper[4813]: I1206 16:04:36.637135 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68cf1c73-04c1-4776-a91d-2189db97e9ed","Type":"ContainerStarted","Data":"6da18f0ee2d1d6e141d273b01e75a7bf57a69862051391d87d8adfeb1061ba54"} Dec 06 16:04:36 crc kubenswrapper[4813]: I1206 16:04:36.656827 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-pn5s8"] Dec 06 16:04:36 crc kubenswrapper[4813]: I1206 16:04:36.662277 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.662268194 podStartE2EDuration="3.662268194s" podCreationTimestamp="2025-12-06 16:04:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:04:36.655500406 +0000 UTC m=+1116.546379982" watchObservedRunningTime="2025-12-06 16:04:36.662268194 +0000 UTC m=+1116.553147770" Dec 06 16:04:36 crc kubenswrapper[4813]: I1206 16:04:36.759308 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:04:37 crc kubenswrapper[4813]: I1206 16:04:37.647537 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ec17345c-3d8e-4981-945c-173fb39aab99","Type":"ContainerStarted","Data":"e0903877b8090615f0e2565f57cf56855081ca315a83fbcd70961586d9e4e5ac"} Dec 06 16:04:37 crc kubenswrapper[4813]: I1206 16:04:37.647781 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ec17345c-3d8e-4981-945c-173fb39aab99","Type":"ContainerStarted","Data":"718eb6eb98eb5eb354aeccf4c8d8c73478c7ac61cb32fd218d4ec1ae016b37d3"} Dec 06 16:04:37 crc kubenswrapper[4813]: I1206 16:04:37.648771 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-pn5s8" event={"ID":"2ff658cc-71bb-4df5-8ba8-1c0979e03d10","Type":"ContainerStarted","Data":"c68f41f338345b8e6e656453f54d2c02662d5b6c8bb967c8550d44968c15b7ae"} Dec 06 16:04:37 crc kubenswrapper[4813]: I1206 16:04:37.650411 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68cf1c73-04c1-4776-a91d-2189db97e9ed","Type":"ContainerStarted","Data":"c544b91c8fcc60ba51f58e444e291c69efad2b3f1fc66d9e3afc1c9bac097cca"} Dec 06 16:04:37 crc kubenswrapper[4813]: I1206 16:04:37.688921 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.688901592 podStartE2EDuration="3.688901592s" podCreationTimestamp="2025-12-06 16:04:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:04:37.677541674 +0000 UTC m=+1117.568421250" watchObservedRunningTime="2025-12-06 16:04:37.688901592 +0000 UTC m=+1117.579781158" Dec 06 16:04:38 crc kubenswrapper[4813]: I1206 16:04:38.727710 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68cf1c73-04c1-4776-a91d-2189db97e9ed","Type":"ContainerStarted","Data":"205bba6683c6e40183f3c06cae7fbf32ec2a6d08f13cf394ea399dfd4bb13efb"} Dec 06 16:04:39 crc kubenswrapper[4813]: I1206 16:04:39.743558 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68cf1c73-04c1-4776-a91d-2189db97e9ed","Type":"ContainerStarted","Data":"98e818332e55dae944df5fe6505b2f5982ab4e4d9df270da69e51de2d0d7e4f9"} Dec 06 16:04:40 crc kubenswrapper[4813]: I1206 16:04:40.753073 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68cf1c73-04c1-4776-a91d-2189db97e9ed","Type":"ContainerStarted","Data":"47cf39c0fc4a890c9761243fc2f686e77fdfb1140cc0bf3db5a17bae6b3bb538"} Dec 06 16:04:40 crc kubenswrapper[4813]: I1206 16:04:40.753240 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="68cf1c73-04c1-4776-a91d-2189db97e9ed" containerName="ceilometer-central-agent" containerID="cri-o://c544b91c8fcc60ba51f58e444e291c69efad2b3f1fc66d9e3afc1c9bac097cca" gracePeriod=30 Dec 06 16:04:40 crc kubenswrapper[4813]: I1206 16:04:40.753670 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="68cf1c73-04c1-4776-a91d-2189db97e9ed" containerName="sg-core" containerID="cri-o://98e818332e55dae944df5fe6505b2f5982ab4e4d9df270da69e51de2d0d7e4f9" gracePeriod=30 Dec 06 16:04:40 crc kubenswrapper[4813]: I1206 16:04:40.753724 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="68cf1c73-04c1-4776-a91d-2189db97e9ed" containerName="ceilometer-notification-agent" containerID="cri-o://205bba6683c6e40183f3c06cae7fbf32ec2a6d08f13cf394ea399dfd4bb13efb" gracePeriod=30 Dec 06 16:04:40 crc kubenswrapper[4813]: I1206 16:04:40.753742 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 16:04:40 crc kubenswrapper[4813]: I1206 16:04:40.753669 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="68cf1c73-04c1-4776-a91d-2189db97e9ed" containerName="proxy-httpd" containerID="cri-o://47cf39c0fc4a890c9761243fc2f686e77fdfb1140cc0bf3db5a17bae6b3bb538" gracePeriod=30 Dec 06 16:04:40 crc kubenswrapper[4813]: I1206 16:04:40.788162 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.274057254 podStartE2EDuration="5.788129677s" podCreationTimestamp="2025-12-06 16:04:35 +0000 UTC" firstStartedPulling="2025-12-06 16:04:36.513564387 +0000 UTC m=+1116.404443963" lastFinishedPulling="2025-12-06 16:04:40.02763681 +0000 UTC m=+1119.918516386" observedRunningTime="2025-12-06 16:04:40.774995852 +0000 UTC m=+1120.665875428" watchObservedRunningTime="2025-12-06 16:04:40.788129677 +0000 UTC m=+1120.679009253" Dec 06 16:04:41 crc kubenswrapper[4813]: I1206 16:04:41.767750 4813 generic.go:334] "Generic (PLEG): container finished" podID="68cf1c73-04c1-4776-a91d-2189db97e9ed" containerID="47cf39c0fc4a890c9761243fc2f686e77fdfb1140cc0bf3db5a17bae6b3bb538" exitCode=0 Dec 06 16:04:41 crc kubenswrapper[4813]: I1206 16:04:41.767776 4813 generic.go:334] "Generic (PLEG): container finished" podID="68cf1c73-04c1-4776-a91d-2189db97e9ed" containerID="98e818332e55dae944df5fe6505b2f5982ab4e4d9df270da69e51de2d0d7e4f9" exitCode=2 Dec 06 16:04:41 crc kubenswrapper[4813]: I1206 16:04:41.767783 4813 generic.go:334] "Generic (PLEG): container finished" podID="68cf1c73-04c1-4776-a91d-2189db97e9ed" containerID="205bba6683c6e40183f3c06cae7fbf32ec2a6d08f13cf394ea399dfd4bb13efb" exitCode=0 Dec 06 16:04:41 crc kubenswrapper[4813]: I1206 16:04:41.767800 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68cf1c73-04c1-4776-a91d-2189db97e9ed","Type":"ContainerDied","Data":"47cf39c0fc4a890c9761243fc2f686e77fdfb1140cc0bf3db5a17bae6b3bb538"} Dec 06 16:04:41 crc kubenswrapper[4813]: I1206 16:04:41.767824 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68cf1c73-04c1-4776-a91d-2189db97e9ed","Type":"ContainerDied","Data":"98e818332e55dae944df5fe6505b2f5982ab4e4d9df270da69e51de2d0d7e4f9"} Dec 06 16:04:41 crc kubenswrapper[4813]: I1206 16:04:41.767834 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68cf1c73-04c1-4776-a91d-2189db97e9ed","Type":"ContainerDied","Data":"205bba6683c6e40183f3c06cae7fbf32ec2a6d08f13cf394ea399dfd4bb13efb"} Dec 06 16:04:43 crc kubenswrapper[4813]: I1206 16:04:43.989298 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 06 16:04:43 crc kubenswrapper[4813]: I1206 16:04:43.989577 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 06 16:04:44 crc kubenswrapper[4813]: I1206 16:04:44.030690 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 06 16:04:44 crc kubenswrapper[4813]: I1206 16:04:44.032354 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 06 16:04:44 crc kubenswrapper[4813]: I1206 16:04:44.806500 4813 generic.go:334] "Generic (PLEG): container finished" podID="68cf1c73-04c1-4776-a91d-2189db97e9ed" containerID="c544b91c8fcc60ba51f58e444e291c69efad2b3f1fc66d9e3afc1c9bac097cca" exitCode=0 Dec 06 16:04:44 crc kubenswrapper[4813]: I1206 16:04:44.806564 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68cf1c73-04c1-4776-a91d-2189db97e9ed","Type":"ContainerDied","Data":"c544b91c8fcc60ba51f58e444e291c69efad2b3f1fc66d9e3afc1c9bac097cca"} Dec 06 16:04:44 crc kubenswrapper[4813]: I1206 16:04:44.806944 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 06 16:04:44 crc kubenswrapper[4813]: I1206 16:04:44.806980 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.002062 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.002116 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.034721 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.048962 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.441577 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.538146 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68cf1c73-04c1-4776-a91d-2189db97e9ed-log-httpd\") pod \"68cf1c73-04c1-4776-a91d-2189db97e9ed\" (UID: \"68cf1c73-04c1-4776-a91d-2189db97e9ed\") " Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.538602 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68cf1c73-04c1-4776-a91d-2189db97e9ed-combined-ca-bundle\") pod \"68cf1c73-04c1-4776-a91d-2189db97e9ed\" (UID: \"68cf1c73-04c1-4776-a91d-2189db97e9ed\") " Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.538684 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68cf1c73-04c1-4776-a91d-2189db97e9ed-scripts\") pod \"68cf1c73-04c1-4776-a91d-2189db97e9ed\" (UID: \"68cf1c73-04c1-4776-a91d-2189db97e9ed\") " Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.538757 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68cf1c73-04c1-4776-a91d-2189db97e9ed-run-httpd\") pod \"68cf1c73-04c1-4776-a91d-2189db97e9ed\" (UID: \"68cf1c73-04c1-4776-a91d-2189db97e9ed\") " Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.538850 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhqjp\" (UniqueName: \"kubernetes.io/projected/68cf1c73-04c1-4776-a91d-2189db97e9ed-kube-api-access-rhqjp\") pod \"68cf1c73-04c1-4776-a91d-2189db97e9ed\" (UID: \"68cf1c73-04c1-4776-a91d-2189db97e9ed\") " Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.538940 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68cf1c73-04c1-4776-a91d-2189db97e9ed-config-data\") pod \"68cf1c73-04c1-4776-a91d-2189db97e9ed\" (UID: \"68cf1c73-04c1-4776-a91d-2189db97e9ed\") " Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.539170 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/68cf1c73-04c1-4776-a91d-2189db97e9ed-sg-core-conf-yaml\") pod \"68cf1c73-04c1-4776-a91d-2189db97e9ed\" (UID: \"68cf1c73-04c1-4776-a91d-2189db97e9ed\") " Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.538861 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68cf1c73-04c1-4776-a91d-2189db97e9ed-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "68cf1c73-04c1-4776-a91d-2189db97e9ed" (UID: "68cf1c73-04c1-4776-a91d-2189db97e9ed"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.539645 4813 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68cf1c73-04c1-4776-a91d-2189db97e9ed-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.539903 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68cf1c73-04c1-4776-a91d-2189db97e9ed-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "68cf1c73-04c1-4776-a91d-2189db97e9ed" (UID: "68cf1c73-04c1-4776-a91d-2189db97e9ed"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.544199 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68cf1c73-04c1-4776-a91d-2189db97e9ed-scripts" (OuterVolumeSpecName: "scripts") pod "68cf1c73-04c1-4776-a91d-2189db97e9ed" (UID: "68cf1c73-04c1-4776-a91d-2189db97e9ed"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.544390 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68cf1c73-04c1-4776-a91d-2189db97e9ed-kube-api-access-rhqjp" (OuterVolumeSpecName: "kube-api-access-rhqjp") pod "68cf1c73-04c1-4776-a91d-2189db97e9ed" (UID: "68cf1c73-04c1-4776-a91d-2189db97e9ed"). InnerVolumeSpecName "kube-api-access-rhqjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.567847 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68cf1c73-04c1-4776-a91d-2189db97e9ed-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "68cf1c73-04c1-4776-a91d-2189db97e9ed" (UID: "68cf1c73-04c1-4776-a91d-2189db97e9ed"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.635464 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68cf1c73-04c1-4776-a91d-2189db97e9ed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "68cf1c73-04c1-4776-a91d-2189db97e9ed" (UID: "68cf1c73-04c1-4776-a91d-2189db97e9ed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.637308 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68cf1c73-04c1-4776-a91d-2189db97e9ed-config-data" (OuterVolumeSpecName: "config-data") pod "68cf1c73-04c1-4776-a91d-2189db97e9ed" (UID: "68cf1c73-04c1-4776-a91d-2189db97e9ed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.642031 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68cf1c73-04c1-4776-a91d-2189db97e9ed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.642052 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68cf1c73-04c1-4776-a91d-2189db97e9ed-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.642062 4813 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68cf1c73-04c1-4776-a91d-2189db97e9ed-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.642072 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhqjp\" (UniqueName: \"kubernetes.io/projected/68cf1c73-04c1-4776-a91d-2189db97e9ed-kube-api-access-rhqjp\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.642085 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68cf1c73-04c1-4776-a91d-2189db97e9ed-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.642096 4813 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/68cf1c73-04c1-4776-a91d-2189db97e9ed-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.815785 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-pn5s8" event={"ID":"2ff658cc-71bb-4df5-8ba8-1c0979e03d10","Type":"ContainerStarted","Data":"9eab219b857b50d7f456aa7914cb1df32c1e15bec7ae7688ec8af024f141900f"} Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.818058 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.818094 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68cf1c73-04c1-4776-a91d-2189db97e9ed","Type":"ContainerDied","Data":"6da18f0ee2d1d6e141d273b01e75a7bf57a69862051391d87d8adfeb1061ba54"} Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.818153 4813 scope.go:117] "RemoveContainer" containerID="47cf39c0fc4a890c9761243fc2f686e77fdfb1140cc0bf3db5a17bae6b3bb538" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.818999 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.819028 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.846523 4813 scope.go:117] "RemoveContainer" containerID="98e818332e55dae944df5fe6505b2f5982ab4e4d9df270da69e51de2d0d7e4f9" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.858961 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-pn5s8" podStartSLOduration=2.29520228 podStartE2EDuration="10.858941235s" podCreationTimestamp="2025-12-06 16:04:35 +0000 UTC" firstStartedPulling="2025-12-06 16:04:36.655024453 +0000 UTC m=+1116.545904029" lastFinishedPulling="2025-12-06 16:04:45.218763408 +0000 UTC m=+1125.109642984" observedRunningTime="2025-12-06 16:04:45.846824727 +0000 UTC m=+1125.737704313" watchObservedRunningTime="2025-12-06 16:04:45.858941235 +0000 UTC m=+1125.749820811" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.882372 4813 scope.go:117] "RemoveContainer" containerID="205bba6683c6e40183f3c06cae7fbf32ec2a6d08f13cf394ea399dfd4bb13efb" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.899880 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.912165 4813 scope.go:117] "RemoveContainer" containerID="c544b91c8fcc60ba51f58e444e291c69efad2b3f1fc66d9e3afc1c9bac097cca" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.917565 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.926913 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:04:45 crc kubenswrapper[4813]: E1206 16:04:45.927246 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68cf1c73-04c1-4776-a91d-2189db97e9ed" containerName="ceilometer-notification-agent" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.927295 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="68cf1c73-04c1-4776-a91d-2189db97e9ed" containerName="ceilometer-notification-agent" Dec 06 16:04:45 crc kubenswrapper[4813]: E1206 16:04:45.927310 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68cf1c73-04c1-4776-a91d-2189db97e9ed" containerName="sg-core" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.927316 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="68cf1c73-04c1-4776-a91d-2189db97e9ed" containerName="sg-core" Dec 06 16:04:45 crc kubenswrapper[4813]: E1206 16:04:45.927326 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68cf1c73-04c1-4776-a91d-2189db97e9ed" containerName="ceilometer-central-agent" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.927332 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="68cf1c73-04c1-4776-a91d-2189db97e9ed" containerName="ceilometer-central-agent" Dec 06 16:04:45 crc kubenswrapper[4813]: E1206 16:04:45.927345 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68cf1c73-04c1-4776-a91d-2189db97e9ed" containerName="proxy-httpd" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.927351 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="68cf1c73-04c1-4776-a91d-2189db97e9ed" containerName="proxy-httpd" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.929661 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="68cf1c73-04c1-4776-a91d-2189db97e9ed" containerName="proxy-httpd" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.929707 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="68cf1c73-04c1-4776-a91d-2189db97e9ed" containerName="ceilometer-notification-agent" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.929732 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="68cf1c73-04c1-4776-a91d-2189db97e9ed" containerName="ceilometer-central-agent" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.929750 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="68cf1c73-04c1-4776-a91d-2189db97e9ed" containerName="sg-core" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.931465 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.936718 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.936912 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 16:04:45 crc kubenswrapper[4813]: I1206 16:04:45.937211 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:04:46 crc kubenswrapper[4813]: I1206 16:04:46.053502 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3860b8fd-910e-472c-af7e-22021b9c5f42-config-data\") pod \"ceilometer-0\" (UID: \"3860b8fd-910e-472c-af7e-22021b9c5f42\") " pod="openstack/ceilometer-0" Dec 06 16:04:46 crc kubenswrapper[4813]: I1206 16:04:46.053796 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3860b8fd-910e-472c-af7e-22021b9c5f42-log-httpd\") pod \"ceilometer-0\" (UID: \"3860b8fd-910e-472c-af7e-22021b9c5f42\") " pod="openstack/ceilometer-0" Dec 06 16:04:46 crc kubenswrapper[4813]: I1206 16:04:46.053842 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3860b8fd-910e-472c-af7e-22021b9c5f42-run-httpd\") pod \"ceilometer-0\" (UID: \"3860b8fd-910e-472c-af7e-22021b9c5f42\") " pod="openstack/ceilometer-0" Dec 06 16:04:46 crc kubenswrapper[4813]: I1206 16:04:46.053867 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3860b8fd-910e-472c-af7e-22021b9c5f42-scripts\") pod \"ceilometer-0\" (UID: \"3860b8fd-910e-472c-af7e-22021b9c5f42\") " pod="openstack/ceilometer-0" Dec 06 16:04:46 crc kubenswrapper[4813]: I1206 16:04:46.053899 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sd55v\" (UniqueName: \"kubernetes.io/projected/3860b8fd-910e-472c-af7e-22021b9c5f42-kube-api-access-sd55v\") pod \"ceilometer-0\" (UID: \"3860b8fd-910e-472c-af7e-22021b9c5f42\") " pod="openstack/ceilometer-0" Dec 06 16:04:46 crc kubenswrapper[4813]: I1206 16:04:46.053920 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3860b8fd-910e-472c-af7e-22021b9c5f42-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3860b8fd-910e-472c-af7e-22021b9c5f42\") " pod="openstack/ceilometer-0" Dec 06 16:04:46 crc kubenswrapper[4813]: I1206 16:04:46.053957 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3860b8fd-910e-472c-af7e-22021b9c5f42-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3860b8fd-910e-472c-af7e-22021b9c5f42\") " pod="openstack/ceilometer-0" Dec 06 16:04:46 crc kubenswrapper[4813]: I1206 16:04:46.155842 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3860b8fd-910e-472c-af7e-22021b9c5f42-config-data\") pod \"ceilometer-0\" (UID: \"3860b8fd-910e-472c-af7e-22021b9c5f42\") " pod="openstack/ceilometer-0" Dec 06 16:04:46 crc kubenswrapper[4813]: I1206 16:04:46.155890 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3860b8fd-910e-472c-af7e-22021b9c5f42-log-httpd\") pod \"ceilometer-0\" (UID: \"3860b8fd-910e-472c-af7e-22021b9c5f42\") " pod="openstack/ceilometer-0" Dec 06 16:04:46 crc kubenswrapper[4813]: I1206 16:04:46.155932 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3860b8fd-910e-472c-af7e-22021b9c5f42-run-httpd\") pod \"ceilometer-0\" (UID: \"3860b8fd-910e-472c-af7e-22021b9c5f42\") " pod="openstack/ceilometer-0" Dec 06 16:04:46 crc kubenswrapper[4813]: I1206 16:04:46.155958 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3860b8fd-910e-472c-af7e-22021b9c5f42-scripts\") pod \"ceilometer-0\" (UID: \"3860b8fd-910e-472c-af7e-22021b9c5f42\") " pod="openstack/ceilometer-0" Dec 06 16:04:46 crc kubenswrapper[4813]: I1206 16:04:46.155988 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sd55v\" (UniqueName: \"kubernetes.io/projected/3860b8fd-910e-472c-af7e-22021b9c5f42-kube-api-access-sd55v\") pod \"ceilometer-0\" (UID: \"3860b8fd-910e-472c-af7e-22021b9c5f42\") " pod="openstack/ceilometer-0" Dec 06 16:04:46 crc kubenswrapper[4813]: I1206 16:04:46.156009 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3860b8fd-910e-472c-af7e-22021b9c5f42-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3860b8fd-910e-472c-af7e-22021b9c5f42\") " pod="openstack/ceilometer-0" Dec 06 16:04:46 crc kubenswrapper[4813]: I1206 16:04:46.156045 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3860b8fd-910e-472c-af7e-22021b9c5f42-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3860b8fd-910e-472c-af7e-22021b9c5f42\") " pod="openstack/ceilometer-0" Dec 06 16:04:46 crc kubenswrapper[4813]: I1206 16:04:46.157270 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3860b8fd-910e-472c-af7e-22021b9c5f42-log-httpd\") pod \"ceilometer-0\" (UID: \"3860b8fd-910e-472c-af7e-22021b9c5f42\") " pod="openstack/ceilometer-0" Dec 06 16:04:46 crc kubenswrapper[4813]: I1206 16:04:46.157534 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3860b8fd-910e-472c-af7e-22021b9c5f42-run-httpd\") pod \"ceilometer-0\" (UID: \"3860b8fd-910e-472c-af7e-22021b9c5f42\") " pod="openstack/ceilometer-0" Dec 06 16:04:46 crc kubenswrapper[4813]: I1206 16:04:46.160602 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3860b8fd-910e-472c-af7e-22021b9c5f42-config-data\") pod \"ceilometer-0\" (UID: \"3860b8fd-910e-472c-af7e-22021b9c5f42\") " pod="openstack/ceilometer-0" Dec 06 16:04:46 crc kubenswrapper[4813]: I1206 16:04:46.160842 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3860b8fd-910e-472c-af7e-22021b9c5f42-scripts\") pod \"ceilometer-0\" (UID: \"3860b8fd-910e-472c-af7e-22021b9c5f42\") " pod="openstack/ceilometer-0" Dec 06 16:04:46 crc kubenswrapper[4813]: I1206 16:04:46.162491 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3860b8fd-910e-472c-af7e-22021b9c5f42-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3860b8fd-910e-472c-af7e-22021b9c5f42\") " pod="openstack/ceilometer-0" Dec 06 16:04:46 crc kubenswrapper[4813]: I1206 16:04:46.163090 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3860b8fd-910e-472c-af7e-22021b9c5f42-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3860b8fd-910e-472c-af7e-22021b9c5f42\") " pod="openstack/ceilometer-0" Dec 06 16:04:46 crc kubenswrapper[4813]: I1206 16:04:46.179000 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sd55v\" (UniqueName: \"kubernetes.io/projected/3860b8fd-910e-472c-af7e-22021b9c5f42-kube-api-access-sd55v\") pod \"ceilometer-0\" (UID: \"3860b8fd-910e-472c-af7e-22021b9c5f42\") " pod="openstack/ceilometer-0" Dec 06 16:04:46 crc kubenswrapper[4813]: I1206 16:04:46.277487 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 16:04:46 crc kubenswrapper[4813]: I1206 16:04:46.496441 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68cf1c73-04c1-4776-a91d-2189db97e9ed" path="/var/lib/kubelet/pods/68cf1c73-04c1-4776-a91d-2189db97e9ed/volumes" Dec 06 16:04:46 crc kubenswrapper[4813]: I1206 16:04:46.776129 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:04:46 crc kubenswrapper[4813]: I1206 16:04:46.827464 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3860b8fd-910e-472c-af7e-22021b9c5f42","Type":"ContainerStarted","Data":"e3340d255397d2ecb2720af5f4b9534bfea2b04ddadbeff378958614dc07f4d8"} Dec 06 16:04:47 crc kubenswrapper[4813]: I1206 16:04:47.041388 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 06 16:04:47 crc kubenswrapper[4813]: I1206 16:04:47.041497 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 16:04:47 crc kubenswrapper[4813]: I1206 16:04:47.055704 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 06 16:04:47 crc kubenswrapper[4813]: I1206 16:04:47.837758 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 16:04:47 crc kubenswrapper[4813]: I1206 16:04:47.838227 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 16:04:47 crc kubenswrapper[4813]: I1206 16:04:47.837771 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3860b8fd-910e-472c-af7e-22021b9c5f42","Type":"ContainerStarted","Data":"2b1742606d409605cb6948ff573d795ff2b4c04a8d7825e8a117f90e873f7472"} Dec 06 16:04:48 crc kubenswrapper[4813]: I1206 16:04:48.092647 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 06 16:04:48 crc kubenswrapper[4813]: I1206 16:04:48.541870 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 06 16:04:48 crc kubenswrapper[4813]: I1206 16:04:48.847965 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3860b8fd-910e-472c-af7e-22021b9c5f42","Type":"ContainerStarted","Data":"ea14b6dcb07649c817b1db8a8ab1402ac31c0fe6aa8e7e0e9b4866d917a94aa8"} Dec 06 16:04:49 crc kubenswrapper[4813]: I1206 16:04:49.855987 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3860b8fd-910e-472c-af7e-22021b9c5f42","Type":"ContainerStarted","Data":"67331f1b42d28f68de93ac94725527090c4590e29a20c3f9e232a471759d8b2b"} Dec 06 16:04:50 crc kubenswrapper[4813]: I1206 16:04:50.866221 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3860b8fd-910e-472c-af7e-22021b9c5f42","Type":"ContainerStarted","Data":"258aee34ba8d549328acc52f2ed0f26867e1376d55703490936bf49c0e56ef3a"} Dec 06 16:04:50 crc kubenswrapper[4813]: I1206 16:04:50.866651 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 16:04:58 crc kubenswrapper[4813]: I1206 16:04:58.962232 4813 generic.go:334] "Generic (PLEG): container finished" podID="2ff658cc-71bb-4df5-8ba8-1c0979e03d10" containerID="9eab219b857b50d7f456aa7914cb1df32c1e15bec7ae7688ec8af024f141900f" exitCode=0 Dec 06 16:04:58 crc kubenswrapper[4813]: I1206 16:04:58.962384 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-pn5s8" event={"ID":"2ff658cc-71bb-4df5-8ba8-1c0979e03d10","Type":"ContainerDied","Data":"9eab219b857b50d7f456aa7914cb1df32c1e15bec7ae7688ec8af024f141900f"} Dec 06 16:04:58 crc kubenswrapper[4813]: I1206 16:04:58.982899 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=10.414566017 podStartE2EDuration="13.982883005s" podCreationTimestamp="2025-12-06 16:04:45 +0000 UTC" firstStartedPulling="2025-12-06 16:04:46.770271266 +0000 UTC m=+1126.661150842" lastFinishedPulling="2025-12-06 16:04:50.338588254 +0000 UTC m=+1130.229467830" observedRunningTime="2025-12-06 16:04:50.904617073 +0000 UTC m=+1130.795496649" watchObservedRunningTime="2025-12-06 16:04:58.982883005 +0000 UTC m=+1138.873762581" Dec 06 16:05:00 crc kubenswrapper[4813]: I1206 16:05:00.375080 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-pn5s8" Dec 06 16:05:00 crc kubenswrapper[4813]: I1206 16:05:00.553509 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ff658cc-71bb-4df5-8ba8-1c0979e03d10-scripts\") pod \"2ff658cc-71bb-4df5-8ba8-1c0979e03d10\" (UID: \"2ff658cc-71bb-4df5-8ba8-1c0979e03d10\") " Dec 06 16:05:00 crc kubenswrapper[4813]: I1206 16:05:00.554776 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbmlq\" (UniqueName: \"kubernetes.io/projected/2ff658cc-71bb-4df5-8ba8-1c0979e03d10-kube-api-access-gbmlq\") pod \"2ff658cc-71bb-4df5-8ba8-1c0979e03d10\" (UID: \"2ff658cc-71bb-4df5-8ba8-1c0979e03d10\") " Dec 06 16:05:00 crc kubenswrapper[4813]: I1206 16:05:00.554862 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ff658cc-71bb-4df5-8ba8-1c0979e03d10-config-data\") pod \"2ff658cc-71bb-4df5-8ba8-1c0979e03d10\" (UID: \"2ff658cc-71bb-4df5-8ba8-1c0979e03d10\") " Dec 06 16:05:00 crc kubenswrapper[4813]: I1206 16:05:00.555008 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ff658cc-71bb-4df5-8ba8-1c0979e03d10-combined-ca-bundle\") pod \"2ff658cc-71bb-4df5-8ba8-1c0979e03d10\" (UID: \"2ff658cc-71bb-4df5-8ba8-1c0979e03d10\") " Dec 06 16:05:00 crc kubenswrapper[4813]: I1206 16:05:00.561494 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ff658cc-71bb-4df5-8ba8-1c0979e03d10-scripts" (OuterVolumeSpecName: "scripts") pod "2ff658cc-71bb-4df5-8ba8-1c0979e03d10" (UID: "2ff658cc-71bb-4df5-8ba8-1c0979e03d10"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:05:00 crc kubenswrapper[4813]: I1206 16:05:00.571649 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ff658cc-71bb-4df5-8ba8-1c0979e03d10-kube-api-access-gbmlq" (OuterVolumeSpecName: "kube-api-access-gbmlq") pod "2ff658cc-71bb-4df5-8ba8-1c0979e03d10" (UID: "2ff658cc-71bb-4df5-8ba8-1c0979e03d10"). InnerVolumeSpecName "kube-api-access-gbmlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:05:00 crc kubenswrapper[4813]: I1206 16:05:00.580400 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ff658cc-71bb-4df5-8ba8-1c0979e03d10-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2ff658cc-71bb-4df5-8ba8-1c0979e03d10" (UID: "2ff658cc-71bb-4df5-8ba8-1c0979e03d10"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:05:00 crc kubenswrapper[4813]: I1206 16:05:00.614443 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ff658cc-71bb-4df5-8ba8-1c0979e03d10-config-data" (OuterVolumeSpecName: "config-data") pod "2ff658cc-71bb-4df5-8ba8-1c0979e03d10" (UID: "2ff658cc-71bb-4df5-8ba8-1c0979e03d10"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:05:00 crc kubenswrapper[4813]: I1206 16:05:00.657241 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ff658cc-71bb-4df5-8ba8-1c0979e03d10-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:00 crc kubenswrapper[4813]: I1206 16:05:00.657300 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ff658cc-71bb-4df5-8ba8-1c0979e03d10-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:00 crc kubenswrapper[4813]: I1206 16:05:00.657317 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ff658cc-71bb-4df5-8ba8-1c0979e03d10-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:00 crc kubenswrapper[4813]: I1206 16:05:00.657329 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbmlq\" (UniqueName: \"kubernetes.io/projected/2ff658cc-71bb-4df5-8ba8-1c0979e03d10-kube-api-access-gbmlq\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:00 crc kubenswrapper[4813]: I1206 16:05:00.988088 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-pn5s8" Dec 06 16:05:00 crc kubenswrapper[4813]: I1206 16:05:00.988078 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-pn5s8" event={"ID":"2ff658cc-71bb-4df5-8ba8-1c0979e03d10","Type":"ContainerDied","Data":"c68f41f338345b8e6e656453f54d2c02662d5b6c8bb967c8550d44968c15b7ae"} Dec 06 16:05:00 crc kubenswrapper[4813]: I1206 16:05:00.988167 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c68f41f338345b8e6e656453f54d2c02662d5b6c8bb967c8550d44968c15b7ae" Dec 06 16:05:01 crc kubenswrapper[4813]: I1206 16:05:01.174100 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 16:05:01 crc kubenswrapper[4813]: E1206 16:05:01.174694 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff658cc-71bb-4df5-8ba8-1c0979e03d10" containerName="nova-cell0-conductor-db-sync" Dec 06 16:05:01 crc kubenswrapper[4813]: I1206 16:05:01.174779 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff658cc-71bb-4df5-8ba8-1c0979e03d10" containerName="nova-cell0-conductor-db-sync" Dec 06 16:05:01 crc kubenswrapper[4813]: I1206 16:05:01.175254 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ff658cc-71bb-4df5-8ba8-1c0979e03d10" containerName="nova-cell0-conductor-db-sync" Dec 06 16:05:01 crc kubenswrapper[4813]: I1206 16:05:01.187319 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 16:05:01 crc kubenswrapper[4813]: I1206 16:05:01.187436 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 06 16:05:01 crc kubenswrapper[4813]: I1206 16:05:01.190077 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-9sldw" Dec 06 16:05:01 crc kubenswrapper[4813]: I1206 16:05:01.190345 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 06 16:05:01 crc kubenswrapper[4813]: I1206 16:05:01.371602 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db38328d-a236-4b36-bb91-9a9b818d39f1-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"db38328d-a236-4b36-bb91-9a9b818d39f1\") " pod="openstack/nova-cell0-conductor-0" Dec 06 16:05:01 crc kubenswrapper[4813]: I1206 16:05:01.372323 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrdfv\" (UniqueName: \"kubernetes.io/projected/db38328d-a236-4b36-bb91-9a9b818d39f1-kube-api-access-xrdfv\") pod \"nova-cell0-conductor-0\" (UID: \"db38328d-a236-4b36-bb91-9a9b818d39f1\") " pod="openstack/nova-cell0-conductor-0" Dec 06 16:05:01 crc kubenswrapper[4813]: I1206 16:05:01.372377 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db38328d-a236-4b36-bb91-9a9b818d39f1-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"db38328d-a236-4b36-bb91-9a9b818d39f1\") " pod="openstack/nova-cell0-conductor-0" Dec 06 16:05:01 crc kubenswrapper[4813]: I1206 16:05:01.474947 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db38328d-a236-4b36-bb91-9a9b818d39f1-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"db38328d-a236-4b36-bb91-9a9b818d39f1\") " pod="openstack/nova-cell0-conductor-0" Dec 06 16:05:01 crc kubenswrapper[4813]: I1206 16:05:01.475209 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrdfv\" (UniqueName: \"kubernetes.io/projected/db38328d-a236-4b36-bb91-9a9b818d39f1-kube-api-access-xrdfv\") pod \"nova-cell0-conductor-0\" (UID: \"db38328d-a236-4b36-bb91-9a9b818d39f1\") " pod="openstack/nova-cell0-conductor-0" Dec 06 16:05:01 crc kubenswrapper[4813]: I1206 16:05:01.475250 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db38328d-a236-4b36-bb91-9a9b818d39f1-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"db38328d-a236-4b36-bb91-9a9b818d39f1\") " pod="openstack/nova-cell0-conductor-0" Dec 06 16:05:01 crc kubenswrapper[4813]: I1206 16:05:01.479464 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db38328d-a236-4b36-bb91-9a9b818d39f1-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"db38328d-a236-4b36-bb91-9a9b818d39f1\") " pod="openstack/nova-cell0-conductor-0" Dec 06 16:05:01 crc kubenswrapper[4813]: I1206 16:05:01.482172 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db38328d-a236-4b36-bb91-9a9b818d39f1-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"db38328d-a236-4b36-bb91-9a9b818d39f1\") " pod="openstack/nova-cell0-conductor-0" Dec 06 16:05:01 crc kubenswrapper[4813]: I1206 16:05:01.498465 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrdfv\" (UniqueName: \"kubernetes.io/projected/db38328d-a236-4b36-bb91-9a9b818d39f1-kube-api-access-xrdfv\") pod \"nova-cell0-conductor-0\" (UID: \"db38328d-a236-4b36-bb91-9a9b818d39f1\") " pod="openstack/nova-cell0-conductor-0" Dec 06 16:05:01 crc kubenswrapper[4813]: I1206 16:05:01.527507 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 06 16:05:02 crc kubenswrapper[4813]: W1206 16:05:02.005008 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb38328d_a236_4b36_bb91_9a9b818d39f1.slice/crio-5117616dfc3f3aa6123dd6c6648d6c9baebb9ef1954a8e61c4a0590a06eadaa2 WatchSource:0}: Error finding container 5117616dfc3f3aa6123dd6c6648d6c9baebb9ef1954a8e61c4a0590a06eadaa2: Status 404 returned error can't find the container with id 5117616dfc3f3aa6123dd6c6648d6c9baebb9ef1954a8e61c4a0590a06eadaa2 Dec 06 16:05:02 crc kubenswrapper[4813]: I1206 16:05:02.007367 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 16:05:03 crc kubenswrapper[4813]: I1206 16:05:03.018915 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"db38328d-a236-4b36-bb91-9a9b818d39f1","Type":"ContainerStarted","Data":"56a96e356cc032ff64e603f4a87d03ba7ca11a12823483ebfa78cae0b3add211"} Dec 06 16:05:03 crc kubenswrapper[4813]: I1206 16:05:03.019392 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"db38328d-a236-4b36-bb91-9a9b818d39f1","Type":"ContainerStarted","Data":"5117616dfc3f3aa6123dd6c6648d6c9baebb9ef1954a8e61c4a0590a06eadaa2"} Dec 06 16:05:03 crc kubenswrapper[4813]: I1206 16:05:03.019422 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 06 16:05:03 crc kubenswrapper[4813]: I1206 16:05:03.059820 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.059756344 podStartE2EDuration="2.059756344s" podCreationTimestamp="2025-12-06 16:05:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:05:03.046735452 +0000 UTC m=+1142.937615038" watchObservedRunningTime="2025-12-06 16:05:03.059756344 +0000 UTC m=+1142.950635950" Dec 06 16:05:11 crc kubenswrapper[4813]: I1206 16:05:11.566748 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.132328 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-xwdwl"] Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.134082 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xwdwl" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.136345 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.145825 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.147115 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-xwdwl"] Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.207036 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fc8a5f4-00b7-434b-b375-74d5a95e3fd1-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-xwdwl\" (UID: \"8fc8a5f4-00b7-434b-b375-74d5a95e3fd1\") " pod="openstack/nova-cell0-cell-mapping-xwdwl" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.207103 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fc8a5f4-00b7-434b-b375-74d5a95e3fd1-config-data\") pod \"nova-cell0-cell-mapping-xwdwl\" (UID: \"8fc8a5f4-00b7-434b-b375-74d5a95e3fd1\") " pod="openstack/nova-cell0-cell-mapping-xwdwl" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.207147 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fc8a5f4-00b7-434b-b375-74d5a95e3fd1-scripts\") pod \"nova-cell0-cell-mapping-xwdwl\" (UID: \"8fc8a5f4-00b7-434b-b375-74d5a95e3fd1\") " pod="openstack/nova-cell0-cell-mapping-xwdwl" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.207201 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r69zj\" (UniqueName: \"kubernetes.io/projected/8fc8a5f4-00b7-434b-b375-74d5a95e3fd1-kube-api-access-r69zj\") pod \"nova-cell0-cell-mapping-xwdwl\" (UID: \"8fc8a5f4-00b7-434b-b375-74d5a95e3fd1\") " pod="openstack/nova-cell0-cell-mapping-xwdwl" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.308056 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fc8a5f4-00b7-434b-b375-74d5a95e3fd1-config-data\") pod \"nova-cell0-cell-mapping-xwdwl\" (UID: \"8fc8a5f4-00b7-434b-b375-74d5a95e3fd1\") " pod="openstack/nova-cell0-cell-mapping-xwdwl" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.308116 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fc8a5f4-00b7-434b-b375-74d5a95e3fd1-scripts\") pod \"nova-cell0-cell-mapping-xwdwl\" (UID: \"8fc8a5f4-00b7-434b-b375-74d5a95e3fd1\") " pod="openstack/nova-cell0-cell-mapping-xwdwl" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.308175 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r69zj\" (UniqueName: \"kubernetes.io/projected/8fc8a5f4-00b7-434b-b375-74d5a95e3fd1-kube-api-access-r69zj\") pod \"nova-cell0-cell-mapping-xwdwl\" (UID: \"8fc8a5f4-00b7-434b-b375-74d5a95e3fd1\") " pod="openstack/nova-cell0-cell-mapping-xwdwl" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.308220 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fc8a5f4-00b7-434b-b375-74d5a95e3fd1-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-xwdwl\" (UID: \"8fc8a5f4-00b7-434b-b375-74d5a95e3fd1\") " pod="openstack/nova-cell0-cell-mapping-xwdwl" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.321013 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fc8a5f4-00b7-434b-b375-74d5a95e3fd1-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-xwdwl\" (UID: \"8fc8a5f4-00b7-434b-b375-74d5a95e3fd1\") " pod="openstack/nova-cell0-cell-mapping-xwdwl" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.326917 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fc8a5f4-00b7-434b-b375-74d5a95e3fd1-scripts\") pod \"nova-cell0-cell-mapping-xwdwl\" (UID: \"8fc8a5f4-00b7-434b-b375-74d5a95e3fd1\") " pod="openstack/nova-cell0-cell-mapping-xwdwl" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.346873 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fc8a5f4-00b7-434b-b375-74d5a95e3fd1-config-data\") pod \"nova-cell0-cell-mapping-xwdwl\" (UID: \"8fc8a5f4-00b7-434b-b375-74d5a95e3fd1\") " pod="openstack/nova-cell0-cell-mapping-xwdwl" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.359648 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r69zj\" (UniqueName: \"kubernetes.io/projected/8fc8a5f4-00b7-434b-b375-74d5a95e3fd1-kube-api-access-r69zj\") pod \"nova-cell0-cell-mapping-xwdwl\" (UID: \"8fc8a5f4-00b7-434b-b375-74d5a95e3fd1\") " pod="openstack/nova-cell0-cell-mapping-xwdwl" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.410980 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.412793 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.426071 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.433616 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.446419 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.447943 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.452099 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.496986 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xwdwl" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.517653 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.523798 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97c18b03-4126-4af3-a790-b7450e54c418-logs\") pod \"nova-api-0\" (UID: \"97c18b03-4126-4af3-a790-b7450e54c418\") " pod="openstack/nova-api-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.523839 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlk6p\" (UniqueName: \"kubernetes.io/projected/8cf165d0-93ba-4070-aa79-d659f77c2ffd-kube-api-access-dlk6p\") pod \"nova-cell1-novncproxy-0\" (UID: \"8cf165d0-93ba-4070-aa79-d659f77c2ffd\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.523864 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97c18b03-4126-4af3-a790-b7450e54c418-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"97c18b03-4126-4af3-a790-b7450e54c418\") " pod="openstack/nova-api-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.523903 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cf165d0-93ba-4070-aa79-d659f77c2ffd-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8cf165d0-93ba-4070-aa79-d659f77c2ffd\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.523923 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cf165d0-93ba-4070-aa79-d659f77c2ffd-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8cf165d0-93ba-4070-aa79-d659f77c2ffd\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.523963 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97c18b03-4126-4af3-a790-b7450e54c418-config-data\") pod \"nova-api-0\" (UID: \"97c18b03-4126-4af3-a790-b7450e54c418\") " pod="openstack/nova-api-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.524021 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nvf4\" (UniqueName: \"kubernetes.io/projected/97c18b03-4126-4af3-a790-b7450e54c418-kube-api-access-8nvf4\") pod \"nova-api-0\" (UID: \"97c18b03-4126-4af3-a790-b7450e54c418\") " pod="openstack/nova-api-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.604688 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.624771 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cf165d0-93ba-4070-aa79-d659f77c2ffd-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8cf165d0-93ba-4070-aa79-d659f77c2ffd\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.624811 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cf165d0-93ba-4070-aa79-d659f77c2ffd-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8cf165d0-93ba-4070-aa79-d659f77c2ffd\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.624856 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97c18b03-4126-4af3-a790-b7450e54c418-config-data\") pod \"nova-api-0\" (UID: \"97c18b03-4126-4af3-a790-b7450e54c418\") " pod="openstack/nova-api-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.624899 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nvf4\" (UniqueName: \"kubernetes.io/projected/97c18b03-4126-4af3-a790-b7450e54c418-kube-api-access-8nvf4\") pod \"nova-api-0\" (UID: \"97c18b03-4126-4af3-a790-b7450e54c418\") " pod="openstack/nova-api-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.624941 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97c18b03-4126-4af3-a790-b7450e54c418-logs\") pod \"nova-api-0\" (UID: \"97c18b03-4126-4af3-a790-b7450e54c418\") " pod="openstack/nova-api-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.624960 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlk6p\" (UniqueName: \"kubernetes.io/projected/8cf165d0-93ba-4070-aa79-d659f77c2ffd-kube-api-access-dlk6p\") pod \"nova-cell1-novncproxy-0\" (UID: \"8cf165d0-93ba-4070-aa79-d659f77c2ffd\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.624982 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97c18b03-4126-4af3-a790-b7450e54c418-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"97c18b03-4126-4af3-a790-b7450e54c418\") " pod="openstack/nova-api-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.628062 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.630721 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97c18b03-4126-4af3-a790-b7450e54c418-logs\") pod \"nova-api-0\" (UID: \"97c18b03-4126-4af3-a790-b7450e54c418\") " pod="openstack/nova-api-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.644521 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.645339 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97c18b03-4126-4af3-a790-b7450e54c418-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"97c18b03-4126-4af3-a790-b7450e54c418\") " pod="openstack/nova-api-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.645494 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cf165d0-93ba-4070-aa79-d659f77c2ffd-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8cf165d0-93ba-4070-aa79-d659f77c2ffd\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.650564 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97c18b03-4126-4af3-a790-b7450e54c418-config-data\") pod \"nova-api-0\" (UID: \"97c18b03-4126-4af3-a790-b7450e54c418\") " pod="openstack/nova-api-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.683287 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cf165d0-93ba-4070-aa79-d659f77c2ffd-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8cf165d0-93ba-4070-aa79-d659f77c2ffd\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.693534 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.714672 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.725008 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.727299 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8926fa21-ffb5-4710-bc3a-b38795709e8e-logs\") pod \"nova-metadata-0\" (UID: \"8926fa21-ffb5-4710-bc3a-b38795709e8e\") " pod="openstack/nova-metadata-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.727338 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4vph\" (UniqueName: \"kubernetes.io/projected/8926fa21-ffb5-4710-bc3a-b38795709e8e-kube-api-access-g4vph\") pod \"nova-metadata-0\" (UID: \"8926fa21-ffb5-4710-bc3a-b38795709e8e\") " pod="openstack/nova-metadata-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.727405 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8926fa21-ffb5-4710-bc3a-b38795709e8e-config-data\") pod \"nova-metadata-0\" (UID: \"8926fa21-ffb5-4710-bc3a-b38795709e8e\") " pod="openstack/nova-metadata-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.727433 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8926fa21-ffb5-4710-bc3a-b38795709e8e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8926fa21-ffb5-4710-bc3a-b38795709e8e\") " pod="openstack/nova-metadata-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.727818 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.742361 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.742812 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nvf4\" (UniqueName: \"kubernetes.io/projected/97c18b03-4126-4af3-a790-b7450e54c418-kube-api-access-8nvf4\") pod \"nova-api-0\" (UID: \"97c18b03-4126-4af3-a790-b7450e54c418\") " pod="openstack/nova-api-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.746487 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlk6p\" (UniqueName: \"kubernetes.io/projected/8cf165d0-93ba-4070-aa79-d659f77c2ffd-kube-api-access-dlk6p\") pod \"nova-cell1-novncproxy-0\" (UID: \"8cf165d0-93ba-4070-aa79-d659f77c2ffd\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.779006 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-4w48r"] Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.780884 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-4w48r" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.781931 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.830307 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4vph\" (UniqueName: \"kubernetes.io/projected/8926fa21-ffb5-4710-bc3a-b38795709e8e-kube-api-access-g4vph\") pod \"nova-metadata-0\" (UID: \"8926fa21-ffb5-4710-bc3a-b38795709e8e\") " pod="openstack/nova-metadata-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.830415 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8926fa21-ffb5-4710-bc3a-b38795709e8e-config-data\") pod \"nova-metadata-0\" (UID: \"8926fa21-ffb5-4710-bc3a-b38795709e8e\") " pod="openstack/nova-metadata-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.830451 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8926fa21-ffb5-4710-bc3a-b38795709e8e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8926fa21-ffb5-4710-bc3a-b38795709e8e\") " pod="openstack/nova-metadata-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.830498 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7da70874-b4ec-4a5a-9061-af3983fd7b88-config-data\") pod \"nova-scheduler-0\" (UID: \"7da70874-b4ec-4a5a-9061-af3983fd7b88\") " pod="openstack/nova-scheduler-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.830528 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxnx9\" (UniqueName: \"kubernetes.io/projected/7da70874-b4ec-4a5a-9061-af3983fd7b88-kube-api-access-rxnx9\") pod \"nova-scheduler-0\" (UID: \"7da70874-b4ec-4a5a-9061-af3983fd7b88\") " pod="openstack/nova-scheduler-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.830555 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7da70874-b4ec-4a5a-9061-af3983fd7b88-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7da70874-b4ec-4a5a-9061-af3983fd7b88\") " pod="openstack/nova-scheduler-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.830605 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8926fa21-ffb5-4710-bc3a-b38795709e8e-logs\") pod \"nova-metadata-0\" (UID: \"8926fa21-ffb5-4710-bc3a-b38795709e8e\") " pod="openstack/nova-metadata-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.834930 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8926fa21-ffb5-4710-bc3a-b38795709e8e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8926fa21-ffb5-4710-bc3a-b38795709e8e\") " pod="openstack/nova-metadata-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.842237 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8926fa21-ffb5-4710-bc3a-b38795709e8e-config-data\") pod \"nova-metadata-0\" (UID: \"8926fa21-ffb5-4710-bc3a-b38795709e8e\") " pod="openstack/nova-metadata-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.845859 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8926fa21-ffb5-4710-bc3a-b38795709e8e-logs\") pod \"nova-metadata-0\" (UID: \"8926fa21-ffb5-4710-bc3a-b38795709e8e\") " pod="openstack/nova-metadata-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.877082 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4vph\" (UniqueName: \"kubernetes.io/projected/8926fa21-ffb5-4710-bc3a-b38795709e8e-kube-api-access-g4vph\") pod \"nova-metadata-0\" (UID: \"8926fa21-ffb5-4710-bc3a-b38795709e8e\") " pod="openstack/nova-metadata-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.921308 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-4w48r"] Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.932332 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-dns-svc\") pod \"dnsmasq-dns-bccf8f775-4w48r\" (UID: \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\") " pod="openstack/dnsmasq-dns-bccf8f775-4w48r" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.932399 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvshl\" (UniqueName: \"kubernetes.io/projected/826c1ef8-3395-4fde-9ef3-fac3d0486f41-kube-api-access-lvshl\") pod \"dnsmasq-dns-bccf8f775-4w48r\" (UID: \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\") " pod="openstack/dnsmasq-dns-bccf8f775-4w48r" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.932488 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7da70874-b4ec-4a5a-9061-af3983fd7b88-config-data\") pod \"nova-scheduler-0\" (UID: \"7da70874-b4ec-4a5a-9061-af3983fd7b88\") " pod="openstack/nova-scheduler-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.932534 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxnx9\" (UniqueName: \"kubernetes.io/projected/7da70874-b4ec-4a5a-9061-af3983fd7b88-kube-api-access-rxnx9\") pod \"nova-scheduler-0\" (UID: \"7da70874-b4ec-4a5a-9061-af3983fd7b88\") " pod="openstack/nova-scheduler-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.932559 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7da70874-b4ec-4a5a-9061-af3983fd7b88-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7da70874-b4ec-4a5a-9061-af3983fd7b88\") " pod="openstack/nova-scheduler-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.932578 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-4w48r\" (UID: \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\") " pod="openstack/dnsmasq-dns-bccf8f775-4w48r" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.932614 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-4w48r\" (UID: \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\") " pod="openstack/dnsmasq-dns-bccf8f775-4w48r" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.932643 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-4w48r\" (UID: \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\") " pod="openstack/dnsmasq-dns-bccf8f775-4w48r" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.932661 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-config\") pod \"dnsmasq-dns-bccf8f775-4w48r\" (UID: \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\") " pod="openstack/dnsmasq-dns-bccf8f775-4w48r" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.935969 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7da70874-b4ec-4a5a-9061-af3983fd7b88-config-data\") pod \"nova-scheduler-0\" (UID: \"7da70874-b4ec-4a5a-9061-af3983fd7b88\") " pod="openstack/nova-scheduler-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.936806 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7da70874-b4ec-4a5a-9061-af3983fd7b88-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7da70874-b4ec-4a5a-9061-af3983fd7b88\") " pod="openstack/nova-scheduler-0" Dec 06 16:05:12 crc kubenswrapper[4813]: I1206 16:05:12.957656 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxnx9\" (UniqueName: \"kubernetes.io/projected/7da70874-b4ec-4a5a-9061-af3983fd7b88-kube-api-access-rxnx9\") pod \"nova-scheduler-0\" (UID: \"7da70874-b4ec-4a5a-9061-af3983fd7b88\") " pod="openstack/nova-scheduler-0" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.033748 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvshl\" (UniqueName: \"kubernetes.io/projected/826c1ef8-3395-4fde-9ef3-fac3d0486f41-kube-api-access-lvshl\") pod \"dnsmasq-dns-bccf8f775-4w48r\" (UID: \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\") " pod="openstack/dnsmasq-dns-bccf8f775-4w48r" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.037020 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-4w48r\" (UID: \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\") " pod="openstack/dnsmasq-dns-bccf8f775-4w48r" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.037706 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-4w48r\" (UID: \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\") " pod="openstack/dnsmasq-dns-bccf8f775-4w48r" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.037866 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-4w48r\" (UID: \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\") " pod="openstack/dnsmasq-dns-bccf8f775-4w48r" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.037894 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-config\") pod \"dnsmasq-dns-bccf8f775-4w48r\" (UID: \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\") " pod="openstack/dnsmasq-dns-bccf8f775-4w48r" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.038118 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-dns-svc\") pod \"dnsmasq-dns-bccf8f775-4w48r\" (UID: \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\") " pod="openstack/dnsmasq-dns-bccf8f775-4w48r" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.038546 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-4w48r\" (UID: \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\") " pod="openstack/dnsmasq-dns-bccf8f775-4w48r" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.038586 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-4w48r\" (UID: \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\") " pod="openstack/dnsmasq-dns-bccf8f775-4w48r" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.039091 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-4w48r\" (UID: \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\") " pod="openstack/dnsmasq-dns-bccf8f775-4w48r" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.040371 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-dns-svc\") pod \"dnsmasq-dns-bccf8f775-4w48r\" (UID: \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\") " pod="openstack/dnsmasq-dns-bccf8f775-4w48r" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.040454 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-config\") pod \"dnsmasq-dns-bccf8f775-4w48r\" (UID: \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\") " pod="openstack/dnsmasq-dns-bccf8f775-4w48r" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.042222 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.044560 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.060814 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvshl\" (UniqueName: \"kubernetes.io/projected/826c1ef8-3395-4fde-9ef3-fac3d0486f41-kube-api-access-lvshl\") pod \"dnsmasq-dns-bccf8f775-4w48r\" (UID: \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\") " pod="openstack/dnsmasq-dns-bccf8f775-4w48r" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.094880 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.160426 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-4w48r" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.287131 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-xwdwl"] Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.339303 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 16:05:13 crc kubenswrapper[4813]: W1206 16:05:13.370017 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cf165d0_93ba_4070_aa79_d659f77c2ffd.slice/crio-d74209a66621bfeda97317d6c56e4a1f05e85252c75d6a6a0856d599371d745b WatchSource:0}: Error finding container d74209a66621bfeda97317d6c56e4a1f05e85252c75d6a6a0856d599371d745b: Status 404 returned error can't find the container with id d74209a66621bfeda97317d6c56e4a1f05e85252c75d6a6a0856d599371d745b Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.682605 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.702854 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.813813 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gv4hp"] Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.818034 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-gv4hp" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.830715 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.832534 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.833446 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gv4hp"] Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.876756 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4nvt\" (UniqueName: \"kubernetes.io/projected/4d9f1a15-0cf8-45e2-b4f4-d15d257d7009-kube-api-access-d4nvt\") pod \"nova-cell1-conductor-db-sync-gv4hp\" (UID: \"4d9f1a15-0cf8-45e2-b4f4-d15d257d7009\") " pod="openstack/nova-cell1-conductor-db-sync-gv4hp" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.876999 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d9f1a15-0cf8-45e2-b4f4-d15d257d7009-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-gv4hp\" (UID: \"4d9f1a15-0cf8-45e2-b4f4-d15d257d7009\") " pod="openstack/nova-cell1-conductor-db-sync-gv4hp" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.877106 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d9f1a15-0cf8-45e2-b4f4-d15d257d7009-scripts\") pod \"nova-cell1-conductor-db-sync-gv4hp\" (UID: \"4d9f1a15-0cf8-45e2-b4f4-d15d257d7009\") " pod="openstack/nova-cell1-conductor-db-sync-gv4hp" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.877205 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d9f1a15-0cf8-45e2-b4f4-d15d257d7009-config-data\") pod \"nova-cell1-conductor-db-sync-gv4hp\" (UID: \"4d9f1a15-0cf8-45e2-b4f4-d15d257d7009\") " pod="openstack/nova-cell1-conductor-db-sync-gv4hp" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.945591 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 16:05:13 crc kubenswrapper[4813]: W1206 16:05:13.959642 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7da70874_b4ec_4a5a_9061_af3983fd7b88.slice/crio-7ad4b9b48589e67f08bd44eda3837a9134e4d428addaf896ccb48425517bbc69 WatchSource:0}: Error finding container 7ad4b9b48589e67f08bd44eda3837a9134e4d428addaf896ccb48425517bbc69: Status 404 returned error can't find the container with id 7ad4b9b48589e67f08bd44eda3837a9134e4d428addaf896ccb48425517bbc69 Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.978714 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4nvt\" (UniqueName: \"kubernetes.io/projected/4d9f1a15-0cf8-45e2-b4f4-d15d257d7009-kube-api-access-d4nvt\") pod \"nova-cell1-conductor-db-sync-gv4hp\" (UID: \"4d9f1a15-0cf8-45e2-b4f4-d15d257d7009\") " pod="openstack/nova-cell1-conductor-db-sync-gv4hp" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.978771 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d9f1a15-0cf8-45e2-b4f4-d15d257d7009-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-gv4hp\" (UID: \"4d9f1a15-0cf8-45e2-b4f4-d15d257d7009\") " pod="openstack/nova-cell1-conductor-db-sync-gv4hp" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.978801 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d9f1a15-0cf8-45e2-b4f4-d15d257d7009-scripts\") pod \"nova-cell1-conductor-db-sync-gv4hp\" (UID: \"4d9f1a15-0cf8-45e2-b4f4-d15d257d7009\") " pod="openstack/nova-cell1-conductor-db-sync-gv4hp" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.978853 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d9f1a15-0cf8-45e2-b4f4-d15d257d7009-config-data\") pod \"nova-cell1-conductor-db-sync-gv4hp\" (UID: \"4d9f1a15-0cf8-45e2-b4f4-d15d257d7009\") " pod="openstack/nova-cell1-conductor-db-sync-gv4hp" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.993121 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d9f1a15-0cf8-45e2-b4f4-d15d257d7009-config-data\") pod \"nova-cell1-conductor-db-sync-gv4hp\" (UID: \"4d9f1a15-0cf8-45e2-b4f4-d15d257d7009\") " pod="openstack/nova-cell1-conductor-db-sync-gv4hp" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.993174 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d9f1a15-0cf8-45e2-b4f4-d15d257d7009-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-gv4hp\" (UID: \"4d9f1a15-0cf8-45e2-b4f4-d15d257d7009\") " pod="openstack/nova-cell1-conductor-db-sync-gv4hp" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.993473 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d9f1a15-0cf8-45e2-b4f4-d15d257d7009-scripts\") pod \"nova-cell1-conductor-db-sync-gv4hp\" (UID: \"4d9f1a15-0cf8-45e2-b4f4-d15d257d7009\") " pod="openstack/nova-cell1-conductor-db-sync-gv4hp" Dec 06 16:05:13 crc kubenswrapper[4813]: I1206 16:05:13.994920 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4nvt\" (UniqueName: \"kubernetes.io/projected/4d9f1a15-0cf8-45e2-b4f4-d15d257d7009-kube-api-access-d4nvt\") pod \"nova-cell1-conductor-db-sync-gv4hp\" (UID: \"4d9f1a15-0cf8-45e2-b4f4-d15d257d7009\") " pod="openstack/nova-cell1-conductor-db-sync-gv4hp" Dec 06 16:05:14 crc kubenswrapper[4813]: I1206 16:05:14.028219 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-4w48r"] Dec 06 16:05:14 crc kubenswrapper[4813]: I1206 16:05:14.152415 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-gv4hp" Dec 06 16:05:14 crc kubenswrapper[4813]: I1206 16:05:14.196369 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-4w48r" event={"ID":"826c1ef8-3395-4fde-9ef3-fac3d0486f41","Type":"ContainerStarted","Data":"d90f7d1b3433ad16104b32ede7b6a9de735b7a1a6394f3f8717e678e9e42d176"} Dec 06 16:05:14 crc kubenswrapper[4813]: I1206 16:05:14.198403 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8926fa21-ffb5-4710-bc3a-b38795709e8e","Type":"ContainerStarted","Data":"0bd6f677313ce706a9b9b923c40ae3621c7f9f4c3c7ca910c0f882b909c2aa6b"} Dec 06 16:05:14 crc kubenswrapper[4813]: I1206 16:05:14.200417 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xwdwl" event={"ID":"8fc8a5f4-00b7-434b-b375-74d5a95e3fd1","Type":"ContainerStarted","Data":"a4537509eb048af72dd28535e54e3ecd586c8ede261a2a308f156c700921caa1"} Dec 06 16:05:14 crc kubenswrapper[4813]: I1206 16:05:14.200441 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xwdwl" event={"ID":"8fc8a5f4-00b7-434b-b375-74d5a95e3fd1","Type":"ContainerStarted","Data":"a9bddc90671279b3d03f13803796c644348acd9afb87e0bee54b0f0685425a07"} Dec 06 16:05:14 crc kubenswrapper[4813]: I1206 16:05:14.219939 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-xwdwl" podStartSLOduration=2.219924856 podStartE2EDuration="2.219924856s" podCreationTimestamp="2025-12-06 16:05:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:05:14.217115213 +0000 UTC m=+1154.107994789" watchObservedRunningTime="2025-12-06 16:05:14.219924856 +0000 UTC m=+1154.110804432" Dec 06 16:05:14 crc kubenswrapper[4813]: I1206 16:05:14.238211 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7da70874-b4ec-4a5a-9061-af3983fd7b88","Type":"ContainerStarted","Data":"7ad4b9b48589e67f08bd44eda3837a9134e4d428addaf896ccb48425517bbc69"} Dec 06 16:05:14 crc kubenswrapper[4813]: I1206 16:05:14.240801 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"97c18b03-4126-4af3-a790-b7450e54c418","Type":"ContainerStarted","Data":"ef16b88288e42da9db6bdcb164fa0b4cefd3dc68b4cd38f775533c557a655fcf"} Dec 06 16:05:14 crc kubenswrapper[4813]: I1206 16:05:14.242065 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8cf165d0-93ba-4070-aa79-d659f77c2ffd","Type":"ContainerStarted","Data":"d74209a66621bfeda97317d6c56e4a1f05e85252c75d6a6a0856d599371d745b"} Dec 06 16:05:14 crc kubenswrapper[4813]: I1206 16:05:14.714557 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gv4hp"] Dec 06 16:05:15 crc kubenswrapper[4813]: I1206 16:05:15.313628 4813 generic.go:334] "Generic (PLEG): container finished" podID="826c1ef8-3395-4fde-9ef3-fac3d0486f41" containerID="f8f601193777045356d0473da46e26c63b720ca68e50e80769ff6583b64a8c07" exitCode=0 Dec 06 16:05:15 crc kubenswrapper[4813]: I1206 16:05:15.314726 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-4w48r" event={"ID":"826c1ef8-3395-4fde-9ef3-fac3d0486f41","Type":"ContainerDied","Data":"f8f601193777045356d0473da46e26c63b720ca68e50e80769ff6583b64a8c07"} Dec 06 16:05:16 crc kubenswrapper[4813]: I1206 16:05:16.287388 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 06 16:05:16 crc kubenswrapper[4813]: I1206 16:05:16.337496 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-gv4hp" event={"ID":"4d9f1a15-0cf8-45e2-b4f4-d15d257d7009","Type":"ContainerStarted","Data":"1f8ab6c54e878d8569462182cd71bc06222ec6f4d98b90bad9a0702f23ed4ea8"} Dec 06 16:05:16 crc kubenswrapper[4813]: I1206 16:05:16.554277 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 16:05:16 crc kubenswrapper[4813]: I1206 16:05:16.586661 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 16:05:18 crc kubenswrapper[4813]: I1206 16:05:18.357421 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-4w48r" event={"ID":"826c1ef8-3395-4fde-9ef3-fac3d0486f41","Type":"ContainerStarted","Data":"75d9e1196f0a65ef60f3a05343344987a95b4755a57be6d5479cf51b6d56ea04"} Dec 06 16:05:19 crc kubenswrapper[4813]: I1206 16:05:19.366944 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8cf165d0-93ba-4070-aa79-d659f77c2ffd","Type":"ContainerStarted","Data":"f0f8043b7bb61d5ebdd814f386b66322a981e2053896852e6bbb9d6e8229b396"} Dec 06 16:05:19 crc kubenswrapper[4813]: I1206 16:05:19.367035 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="8cf165d0-93ba-4070-aa79-d659f77c2ffd" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://f0f8043b7bb61d5ebdd814f386b66322a981e2053896852e6bbb9d6e8229b396" gracePeriod=30 Dec 06 16:05:19 crc kubenswrapper[4813]: I1206 16:05:19.370152 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8926fa21-ffb5-4710-bc3a-b38795709e8e","Type":"ContainerStarted","Data":"82d9b9fe4e9bd0dd13e2dd872ffca27f96aa14278886cf46793bbb06d35ff1d3"} Dec 06 16:05:19 crc kubenswrapper[4813]: I1206 16:05:19.370194 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8926fa21-ffb5-4710-bc3a-b38795709e8e","Type":"ContainerStarted","Data":"8391a476663ec29e6502285e788214908680f9c4cfbc6fe58dbd523f178235d9"} Dec 06 16:05:19 crc kubenswrapper[4813]: I1206 16:05:19.370422 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8926fa21-ffb5-4710-bc3a-b38795709e8e" containerName="nova-metadata-metadata" containerID="cri-o://82d9b9fe4e9bd0dd13e2dd872ffca27f96aa14278886cf46793bbb06d35ff1d3" gracePeriod=30 Dec 06 16:05:19 crc kubenswrapper[4813]: I1206 16:05:19.370355 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8926fa21-ffb5-4710-bc3a-b38795709e8e" containerName="nova-metadata-log" containerID="cri-o://8391a476663ec29e6502285e788214908680f9c4cfbc6fe58dbd523f178235d9" gracePeriod=30 Dec 06 16:05:19 crc kubenswrapper[4813]: I1206 16:05:19.375864 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-gv4hp" event={"ID":"4d9f1a15-0cf8-45e2-b4f4-d15d257d7009","Type":"ContainerStarted","Data":"bd8568faea71fd736f6a65c700cae646473782de03144171c2d8ade2038b5179"} Dec 06 16:05:19 crc kubenswrapper[4813]: I1206 16:05:19.382813 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7da70874-b4ec-4a5a-9061-af3983fd7b88","Type":"ContainerStarted","Data":"2b1588c231aa5871f88c2b3c13a462416c153666299d3a05e6367d79b9e062ef"} Dec 06 16:05:19 crc kubenswrapper[4813]: I1206 16:05:19.406424 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"97c18b03-4126-4af3-a790-b7450e54c418","Type":"ContainerStarted","Data":"56a4caccafc3b2b9f412eb23ce642e5c300d84069300ba7fa0b382fefe96ef39"} Dec 06 16:05:19 crc kubenswrapper[4813]: I1206 16:05:19.406468 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"97c18b03-4126-4af3-a790-b7450e54c418","Type":"ContainerStarted","Data":"1fb6e6ef65034b4e720bedb2ef2b57091bf7d57b3931fb0bac81fbfdf396cb1c"} Dec 06 16:05:19 crc kubenswrapper[4813]: I1206 16:05:19.406499 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bccf8f775-4w48r" Dec 06 16:05:19 crc kubenswrapper[4813]: I1206 16:05:19.427177 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:05:19 crc kubenswrapper[4813]: I1206 16:05:19.427231 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:05:19 crc kubenswrapper[4813]: I1206 16:05:19.431376 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.932998718 podStartE2EDuration="7.431356288s" podCreationTimestamp="2025-12-06 16:05:12 +0000 UTC" firstStartedPulling="2025-12-06 16:05:13.393152017 +0000 UTC m=+1153.284031593" lastFinishedPulling="2025-12-06 16:05:17.891509547 +0000 UTC m=+1157.782389163" observedRunningTime="2025-12-06 16:05:19.403839016 +0000 UTC m=+1159.294718592" watchObservedRunningTime="2025-12-06 16:05:19.431356288 +0000 UTC m=+1159.322235864" Dec 06 16:05:19 crc kubenswrapper[4813]: I1206 16:05:19.474761 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.191450628 podStartE2EDuration="7.474744948s" podCreationTimestamp="2025-12-06 16:05:12 +0000 UTC" firstStartedPulling="2025-12-06 16:05:13.702061162 +0000 UTC m=+1153.592940738" lastFinishedPulling="2025-12-06 16:05:17.985355482 +0000 UTC m=+1157.876235058" observedRunningTime="2025-12-06 16:05:19.462727733 +0000 UTC m=+1159.353607309" watchObservedRunningTime="2025-12-06 16:05:19.474744948 +0000 UTC m=+1159.365624524" Dec 06 16:05:19 crc kubenswrapper[4813]: I1206 16:05:19.480178 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-gv4hp" podStartSLOduration=6.480163051 podStartE2EDuration="6.480163051s" podCreationTimestamp="2025-12-06 16:05:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:05:19.429029717 +0000 UTC m=+1159.319909283" watchObservedRunningTime="2025-12-06 16:05:19.480163051 +0000 UTC m=+1159.371042627" Dec 06 16:05:19 crc kubenswrapper[4813]: I1206 16:05:19.505052 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.532319053 podStartE2EDuration="7.505020584s" podCreationTimestamp="2025-12-06 16:05:12 +0000 UTC" firstStartedPulling="2025-12-06 16:05:13.961629971 +0000 UTC m=+1153.852509547" lastFinishedPulling="2025-12-06 16:05:17.934331502 +0000 UTC m=+1157.825211078" observedRunningTime="2025-12-06 16:05:19.499483628 +0000 UTC m=+1159.390363204" watchObservedRunningTime="2025-12-06 16:05:19.505020584 +0000 UTC m=+1159.395900160" Dec 06 16:05:19 crc kubenswrapper[4813]: I1206 16:05:19.598547 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bccf8f775-4w48r" podStartSLOduration=7.59853222 podStartE2EDuration="7.59853222s" podCreationTimestamp="2025-12-06 16:05:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:05:19.556901187 +0000 UTC m=+1159.447780753" watchObservedRunningTime="2025-12-06 16:05:19.59853222 +0000 UTC m=+1159.489411786" Dec 06 16:05:19 crc kubenswrapper[4813]: I1206 16:05:19.600730 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.343274056 podStartE2EDuration="7.600726298s" podCreationTimestamp="2025-12-06 16:05:12 +0000 UTC" firstStartedPulling="2025-12-06 16:05:13.718935855 +0000 UTC m=+1153.609815431" lastFinishedPulling="2025-12-06 16:05:17.976388097 +0000 UTC m=+1157.867267673" observedRunningTime="2025-12-06 16:05:19.598368776 +0000 UTC m=+1159.489248352" watchObservedRunningTime="2025-12-06 16:05:19.600726298 +0000 UTC m=+1159.491605874" Dec 06 16:05:20 crc kubenswrapper[4813]: I1206 16:05:20.419729 4813 generic.go:334] "Generic (PLEG): container finished" podID="8926fa21-ffb5-4710-bc3a-b38795709e8e" containerID="8391a476663ec29e6502285e788214908680f9c4cfbc6fe58dbd523f178235d9" exitCode=143 Dec 06 16:05:20 crc kubenswrapper[4813]: I1206 16:05:20.419823 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8926fa21-ffb5-4710-bc3a-b38795709e8e","Type":"ContainerDied","Data":"8391a476663ec29e6502285e788214908680f9c4cfbc6fe58dbd523f178235d9"} Dec 06 16:05:21 crc kubenswrapper[4813]: I1206 16:05:21.437688 4813 generic.go:334] "Generic (PLEG): container finished" podID="8926fa21-ffb5-4710-bc3a-b38795709e8e" containerID="82d9b9fe4e9bd0dd13e2dd872ffca27f96aa14278886cf46793bbb06d35ff1d3" exitCode=0 Dec 06 16:05:21 crc kubenswrapper[4813]: I1206 16:05:21.437987 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8926fa21-ffb5-4710-bc3a-b38795709e8e","Type":"ContainerDied","Data":"82d9b9fe4e9bd0dd13e2dd872ffca27f96aa14278886cf46793bbb06d35ff1d3"} Dec 06 16:05:21 crc kubenswrapper[4813]: I1206 16:05:21.687105 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 16:05:21 crc kubenswrapper[4813]: I1206 16:05:21.862146 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8926fa21-ffb5-4710-bc3a-b38795709e8e-combined-ca-bundle\") pod \"8926fa21-ffb5-4710-bc3a-b38795709e8e\" (UID: \"8926fa21-ffb5-4710-bc3a-b38795709e8e\") " Dec 06 16:05:21 crc kubenswrapper[4813]: I1206 16:05:21.863008 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4vph\" (UniqueName: \"kubernetes.io/projected/8926fa21-ffb5-4710-bc3a-b38795709e8e-kube-api-access-g4vph\") pod \"8926fa21-ffb5-4710-bc3a-b38795709e8e\" (UID: \"8926fa21-ffb5-4710-bc3a-b38795709e8e\") " Dec 06 16:05:21 crc kubenswrapper[4813]: I1206 16:05:21.863188 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8926fa21-ffb5-4710-bc3a-b38795709e8e-logs\") pod \"8926fa21-ffb5-4710-bc3a-b38795709e8e\" (UID: \"8926fa21-ffb5-4710-bc3a-b38795709e8e\") " Dec 06 16:05:21 crc kubenswrapper[4813]: I1206 16:05:21.863352 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8926fa21-ffb5-4710-bc3a-b38795709e8e-config-data\") pod \"8926fa21-ffb5-4710-bc3a-b38795709e8e\" (UID: \"8926fa21-ffb5-4710-bc3a-b38795709e8e\") " Dec 06 16:05:21 crc kubenswrapper[4813]: I1206 16:05:21.863605 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8926fa21-ffb5-4710-bc3a-b38795709e8e-logs" (OuterVolumeSpecName: "logs") pod "8926fa21-ffb5-4710-bc3a-b38795709e8e" (UID: "8926fa21-ffb5-4710-bc3a-b38795709e8e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:05:21 crc kubenswrapper[4813]: I1206 16:05:21.864095 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8926fa21-ffb5-4710-bc3a-b38795709e8e-logs\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:21 crc kubenswrapper[4813]: I1206 16:05:21.874548 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8926fa21-ffb5-4710-bc3a-b38795709e8e-kube-api-access-g4vph" (OuterVolumeSpecName: "kube-api-access-g4vph") pod "8926fa21-ffb5-4710-bc3a-b38795709e8e" (UID: "8926fa21-ffb5-4710-bc3a-b38795709e8e"). InnerVolumeSpecName "kube-api-access-g4vph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:05:21 crc kubenswrapper[4813]: I1206 16:05:21.897391 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8926fa21-ffb5-4710-bc3a-b38795709e8e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8926fa21-ffb5-4710-bc3a-b38795709e8e" (UID: "8926fa21-ffb5-4710-bc3a-b38795709e8e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:05:21 crc kubenswrapper[4813]: I1206 16:05:21.911013 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8926fa21-ffb5-4710-bc3a-b38795709e8e-config-data" (OuterVolumeSpecName: "config-data") pod "8926fa21-ffb5-4710-bc3a-b38795709e8e" (UID: "8926fa21-ffb5-4710-bc3a-b38795709e8e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:05:21 crc kubenswrapper[4813]: I1206 16:05:21.967116 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8926fa21-ffb5-4710-bc3a-b38795709e8e-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:21 crc kubenswrapper[4813]: I1206 16:05:21.967161 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8926fa21-ffb5-4710-bc3a-b38795709e8e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:21 crc kubenswrapper[4813]: I1206 16:05:21.967174 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4vph\" (UniqueName: \"kubernetes.io/projected/8926fa21-ffb5-4710-bc3a-b38795709e8e-kube-api-access-g4vph\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.447328 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8926fa21-ffb5-4710-bc3a-b38795709e8e","Type":"ContainerDied","Data":"0bd6f677313ce706a9b9b923c40ae3621c7f9f4c3c7ca910c0f882b909c2aa6b"} Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.447372 4813 scope.go:117] "RemoveContainer" containerID="82d9b9fe4e9bd0dd13e2dd872ffca27f96aa14278886cf46793bbb06d35ff1d3" Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.447475 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.500580 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.510591 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.524667 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 06 16:05:22 crc kubenswrapper[4813]: E1206 16:05:22.525210 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8926fa21-ffb5-4710-bc3a-b38795709e8e" containerName="nova-metadata-log" Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.525236 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="8926fa21-ffb5-4710-bc3a-b38795709e8e" containerName="nova-metadata-log" Dec 06 16:05:22 crc kubenswrapper[4813]: E1206 16:05:22.525355 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8926fa21-ffb5-4710-bc3a-b38795709e8e" containerName="nova-metadata-metadata" Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.525366 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="8926fa21-ffb5-4710-bc3a-b38795709e8e" containerName="nova-metadata-metadata" Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.525649 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="8926fa21-ffb5-4710-bc3a-b38795709e8e" containerName="nova-metadata-metadata" Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.525686 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="8926fa21-ffb5-4710-bc3a-b38795709e8e" containerName="nova-metadata-log" Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.526928 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.530014 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.530325 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.588404 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.591767 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f70a63c7-a091-4b73-b86e-805e3529a69b-logs\") pod \"nova-metadata-0\" (UID: \"f70a63c7-a091-4b73-b86e-805e3529a69b\") " pod="openstack/nova-metadata-0" Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.591821 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttlm8\" (UniqueName: \"kubernetes.io/projected/f70a63c7-a091-4b73-b86e-805e3529a69b-kube-api-access-ttlm8\") pod \"nova-metadata-0\" (UID: \"f70a63c7-a091-4b73-b86e-805e3529a69b\") " pod="openstack/nova-metadata-0" Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.594610 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f70a63c7-a091-4b73-b86e-805e3529a69b-config-data\") pod \"nova-metadata-0\" (UID: \"f70a63c7-a091-4b73-b86e-805e3529a69b\") " pod="openstack/nova-metadata-0" Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.594713 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f70a63c7-a091-4b73-b86e-805e3529a69b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f70a63c7-a091-4b73-b86e-805e3529a69b\") " pod="openstack/nova-metadata-0" Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.595332 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f70a63c7-a091-4b73-b86e-805e3529a69b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f70a63c7-a091-4b73-b86e-805e3529a69b\") " pod="openstack/nova-metadata-0" Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.614578 4813 scope.go:117] "RemoveContainer" containerID="8391a476663ec29e6502285e788214908680f9c4cfbc6fe58dbd523f178235d9" Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.696469 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f70a63c7-a091-4b73-b86e-805e3529a69b-config-data\") pod \"nova-metadata-0\" (UID: \"f70a63c7-a091-4b73-b86e-805e3529a69b\") " pod="openstack/nova-metadata-0" Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.696539 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f70a63c7-a091-4b73-b86e-805e3529a69b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f70a63c7-a091-4b73-b86e-805e3529a69b\") " pod="openstack/nova-metadata-0" Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.696616 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f70a63c7-a091-4b73-b86e-805e3529a69b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f70a63c7-a091-4b73-b86e-805e3529a69b\") " pod="openstack/nova-metadata-0" Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.696684 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f70a63c7-a091-4b73-b86e-805e3529a69b-logs\") pod \"nova-metadata-0\" (UID: \"f70a63c7-a091-4b73-b86e-805e3529a69b\") " pod="openstack/nova-metadata-0" Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.696712 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttlm8\" (UniqueName: \"kubernetes.io/projected/f70a63c7-a091-4b73-b86e-805e3529a69b-kube-api-access-ttlm8\") pod \"nova-metadata-0\" (UID: \"f70a63c7-a091-4b73-b86e-805e3529a69b\") " pod="openstack/nova-metadata-0" Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.697371 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f70a63c7-a091-4b73-b86e-805e3529a69b-logs\") pod \"nova-metadata-0\" (UID: \"f70a63c7-a091-4b73-b86e-805e3529a69b\") " pod="openstack/nova-metadata-0" Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.702943 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f70a63c7-a091-4b73-b86e-805e3529a69b-config-data\") pod \"nova-metadata-0\" (UID: \"f70a63c7-a091-4b73-b86e-805e3529a69b\") " pod="openstack/nova-metadata-0" Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.703121 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f70a63c7-a091-4b73-b86e-805e3529a69b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f70a63c7-a091-4b73-b86e-805e3529a69b\") " pod="openstack/nova-metadata-0" Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.703475 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f70a63c7-a091-4b73-b86e-805e3529a69b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f70a63c7-a091-4b73-b86e-805e3529a69b\") " pod="openstack/nova-metadata-0" Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.721251 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttlm8\" (UniqueName: \"kubernetes.io/projected/f70a63c7-a091-4b73-b86e-805e3529a69b-kube-api-access-ttlm8\") pod \"nova-metadata-0\" (UID: \"f70a63c7-a091-4b73-b86e-805e3529a69b\") " pod="openstack/nova-metadata-0" Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.776191 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.776441 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="c9961e9d-0fa2-4619-9e19-9b2dfd4af40c" containerName="kube-state-metrics" containerID="cri-o://3f45f463a3c6b69a17b6668feff45caeb8fe7254fc4c2843ae755224b30b8589" gracePeriod=30 Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.790450 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:22 crc kubenswrapper[4813]: I1206 16:05:22.853854 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 16:05:23 crc kubenswrapper[4813]: I1206 16:05:23.043224 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 16:05:23 crc kubenswrapper[4813]: I1206 16:05:23.043478 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 16:05:23 crc kubenswrapper[4813]: I1206 16:05:23.098112 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 06 16:05:23 crc kubenswrapper[4813]: I1206 16:05:23.098169 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 06 16:05:23 crc kubenswrapper[4813]: I1206 16:05:23.165461 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bccf8f775-4w48r" Dec 06 16:05:23 crc kubenswrapper[4813]: I1206 16:05:23.169901 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 06 16:05:23 crc kubenswrapper[4813]: I1206 16:05:23.247336 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-j49d2"] Dec 06 16:05:23 crc kubenswrapper[4813]: I1206 16:05:23.247551 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6578955fd5-j49d2" podUID="d7e7e469-ed0d-463c-839c-95427c632d5b" containerName="dnsmasq-dns" containerID="cri-o://f1184bf09f3b6bba16396cf9b3e4d07ce090b43676ce244963ac2cd8f8eddd31" gracePeriod=10 Dec 06 16:05:23 crc kubenswrapper[4813]: I1206 16:05:23.511408 4813 generic.go:334] "Generic (PLEG): container finished" podID="d7e7e469-ed0d-463c-839c-95427c632d5b" containerID="f1184bf09f3b6bba16396cf9b3e4d07ce090b43676ce244963ac2cd8f8eddd31" exitCode=0 Dec 06 16:05:23 crc kubenswrapper[4813]: I1206 16:05:23.511481 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-j49d2" event={"ID":"d7e7e469-ed0d-463c-839c-95427c632d5b","Type":"ContainerDied","Data":"f1184bf09f3b6bba16396cf9b3e4d07ce090b43676ce244963ac2cd8f8eddd31"} Dec 06 16:05:23 crc kubenswrapper[4813]: I1206 16:05:23.530533 4813 generic.go:334] "Generic (PLEG): container finished" podID="c9961e9d-0fa2-4619-9e19-9b2dfd4af40c" containerID="3f45f463a3c6b69a17b6668feff45caeb8fe7254fc4c2843ae755224b30b8589" exitCode=2 Dec 06 16:05:23 crc kubenswrapper[4813]: I1206 16:05:23.531484 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 16:05:23 crc kubenswrapper[4813]: I1206 16:05:23.531849 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c9961e9d-0fa2-4619-9e19-9b2dfd4af40c","Type":"ContainerDied","Data":"3f45f463a3c6b69a17b6668feff45caeb8fe7254fc4c2843ae755224b30b8589"} Dec 06 16:05:23 crc kubenswrapper[4813]: I1206 16:05:23.531886 4813 scope.go:117] "RemoveContainer" containerID="3f45f463a3c6b69a17b6668feff45caeb8fe7254fc4c2843ae755224b30b8589" Dec 06 16:05:23 crc kubenswrapper[4813]: I1206 16:05:23.543092 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wt7cv\" (UniqueName: \"kubernetes.io/projected/c9961e9d-0fa2-4619-9e19-9b2dfd4af40c-kube-api-access-wt7cv\") pod \"c9961e9d-0fa2-4619-9e19-9b2dfd4af40c\" (UID: \"c9961e9d-0fa2-4619-9e19-9b2dfd4af40c\") " Dec 06 16:05:23 crc kubenswrapper[4813]: I1206 16:05:23.551443 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9961e9d-0fa2-4619-9e19-9b2dfd4af40c-kube-api-access-wt7cv" (OuterVolumeSpecName: "kube-api-access-wt7cv") pod "c9961e9d-0fa2-4619-9e19-9b2dfd4af40c" (UID: "c9961e9d-0fa2-4619-9e19-9b2dfd4af40c"). InnerVolumeSpecName "kube-api-access-wt7cv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:05:23 crc kubenswrapper[4813]: I1206 16:05:23.599606 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 06 16:05:23 crc kubenswrapper[4813]: I1206 16:05:23.645658 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wt7cv\" (UniqueName: \"kubernetes.io/projected/c9961e9d-0fa2-4619-9e19-9b2dfd4af40c-kube-api-access-wt7cv\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:23 crc kubenswrapper[4813]: I1206 16:05:23.662866 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 16:05:23 crc kubenswrapper[4813]: I1206 16:05:23.851841 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-j49d2" Dec 06 16:05:23 crc kubenswrapper[4813]: I1206 16:05:23.950672 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rj9bs\" (UniqueName: \"kubernetes.io/projected/d7e7e469-ed0d-463c-839c-95427c632d5b-kube-api-access-rj9bs\") pod \"d7e7e469-ed0d-463c-839c-95427c632d5b\" (UID: \"d7e7e469-ed0d-463c-839c-95427c632d5b\") " Dec 06 16:05:23 crc kubenswrapper[4813]: I1206 16:05:23.950735 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7e7e469-ed0d-463c-839c-95427c632d5b-ovsdbserver-nb\") pod \"d7e7e469-ed0d-463c-839c-95427c632d5b\" (UID: \"d7e7e469-ed0d-463c-839c-95427c632d5b\") " Dec 06 16:05:23 crc kubenswrapper[4813]: I1206 16:05:23.950762 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7e7e469-ed0d-463c-839c-95427c632d5b-config\") pod \"d7e7e469-ed0d-463c-839c-95427c632d5b\" (UID: \"d7e7e469-ed0d-463c-839c-95427c632d5b\") " Dec 06 16:05:23 crc kubenswrapper[4813]: I1206 16:05:23.950838 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7e7e469-ed0d-463c-839c-95427c632d5b-dns-svc\") pod \"d7e7e469-ed0d-463c-839c-95427c632d5b\" (UID: \"d7e7e469-ed0d-463c-839c-95427c632d5b\") " Dec 06 16:05:23 crc kubenswrapper[4813]: I1206 16:05:23.950978 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7e7e469-ed0d-463c-839c-95427c632d5b-dns-swift-storage-0\") pod \"d7e7e469-ed0d-463c-839c-95427c632d5b\" (UID: \"d7e7e469-ed0d-463c-839c-95427c632d5b\") " Dec 06 16:05:23 crc kubenswrapper[4813]: I1206 16:05:23.951009 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7e7e469-ed0d-463c-839c-95427c632d5b-ovsdbserver-sb\") pod \"d7e7e469-ed0d-463c-839c-95427c632d5b\" (UID: \"d7e7e469-ed0d-463c-839c-95427c632d5b\") " Dec 06 16:05:23 crc kubenswrapper[4813]: I1206 16:05:23.971440 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7e7e469-ed0d-463c-839c-95427c632d5b-kube-api-access-rj9bs" (OuterVolumeSpecName: "kube-api-access-rj9bs") pod "d7e7e469-ed0d-463c-839c-95427c632d5b" (UID: "d7e7e469-ed0d-463c-839c-95427c632d5b"). InnerVolumeSpecName "kube-api-access-rj9bs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.053478 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rj9bs\" (UniqueName: \"kubernetes.io/projected/d7e7e469-ed0d-463c-839c-95427c632d5b-kube-api-access-rj9bs\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.060171 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7e7e469-ed0d-463c-839c-95427c632d5b-config" (OuterVolumeSpecName: "config") pod "d7e7e469-ed0d-463c-839c-95427c632d5b" (UID: "d7e7e469-ed0d-463c-839c-95427c632d5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.071898 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7e7e469-ed0d-463c-839c-95427c632d5b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d7e7e469-ed0d-463c-839c-95427c632d5b" (UID: "d7e7e469-ed0d-463c-839c-95427c632d5b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.080999 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7e7e469-ed0d-463c-839c-95427c632d5b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d7e7e469-ed0d-463c-839c-95427c632d5b" (UID: "d7e7e469-ed0d-463c-839c-95427c632d5b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.108814 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7e7e469-ed0d-463c-839c-95427c632d5b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d7e7e469-ed0d-463c-839c-95427c632d5b" (UID: "d7e7e469-ed0d-463c-839c-95427c632d5b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.137470 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="97c18b03-4126-4af3-a790-b7450e54c418" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.185:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.137742 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="97c18b03-4126-4af3-a790-b7450e54c418" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.185:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.154741 4813 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7e7e469-ed0d-463c-839c-95427c632d5b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.154765 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7e7e469-ed0d-463c-839c-95427c632d5b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.154775 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7e7e469-ed0d-463c-839c-95427c632d5b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.154785 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7e7e469-ed0d-463c-839c-95427c632d5b-config\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.163004 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7e7e469-ed0d-463c-839c-95427c632d5b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d7e7e469-ed0d-463c-839c-95427c632d5b" (UID: "d7e7e469-ed0d-463c-839c-95427c632d5b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.256499 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7e7e469-ed0d-463c-839c-95427c632d5b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.495768 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8926fa21-ffb5-4710-bc3a-b38795709e8e" path="/var/lib/kubelet/pods/8926fa21-ffb5-4710-bc3a-b38795709e8e/volumes" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.540961 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-j49d2" event={"ID":"d7e7e469-ed0d-463c-839c-95427c632d5b","Type":"ContainerDied","Data":"a58a45b0f7b1e35e62105d2e77c5e9fb0706feab2dddd9ec0990f6066eb0c8e9"} Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.541009 4813 scope.go:117] "RemoveContainer" containerID="f1184bf09f3b6bba16396cf9b3e4d07ce090b43676ce244963ac2cd8f8eddd31" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.541107 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-j49d2" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.545538 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.545533 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c9961e9d-0fa2-4619-9e19-9b2dfd4af40c","Type":"ContainerDied","Data":"6d6e3a94cb9511ce35608cbdc770d2eed67613433e2e2b89e2eb113161b936d4"} Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.547963 4813 generic.go:334] "Generic (PLEG): container finished" podID="8fc8a5f4-00b7-434b-b375-74d5a95e3fd1" containerID="a4537509eb048af72dd28535e54e3ecd586c8ede261a2a308f156c700921caa1" exitCode=0 Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.548009 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xwdwl" event={"ID":"8fc8a5f4-00b7-434b-b375-74d5a95e3fd1","Type":"ContainerDied","Data":"a4537509eb048af72dd28535e54e3ecd586c8ede261a2a308f156c700921caa1"} Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.554032 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f70a63c7-a091-4b73-b86e-805e3529a69b","Type":"ContainerStarted","Data":"0d54482c9adab1a07bb808250d0b1b6f7d31ce1170140ee8609b9f217dd7333e"} Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.554060 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f70a63c7-a091-4b73-b86e-805e3529a69b","Type":"ContainerStarted","Data":"e429a1d4e041b9aa8abe15788220b63145b20aff70484a7775f4259b265e456a"} Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.554072 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f70a63c7-a091-4b73-b86e-805e3529a69b","Type":"ContainerStarted","Data":"54de8c8e2d8e340e24aff8768c4dbf080cc4293d1fb4b0502bf322a5b7b142ed"} Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.569159 4813 scope.go:117] "RemoveContainer" containerID="ced8f07e40e0ca4da9ceb3d788b88b7be92cd599726686b994d7eb1f2895313a" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.594604 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-j49d2"] Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.594658 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-j49d2"] Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.640644 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.649025 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.673770 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 16:05:24 crc kubenswrapper[4813]: E1206 16:05:24.674168 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7e7e469-ed0d-463c-839c-95427c632d5b" containerName="dnsmasq-dns" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.674186 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7e7e469-ed0d-463c-839c-95427c632d5b" containerName="dnsmasq-dns" Dec 06 16:05:24 crc kubenswrapper[4813]: E1206 16:05:24.674202 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7e7e469-ed0d-463c-839c-95427c632d5b" containerName="init" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.674211 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7e7e469-ed0d-463c-839c-95427c632d5b" containerName="init" Dec 06 16:05:24 crc kubenswrapper[4813]: E1206 16:05:24.674234 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9961e9d-0fa2-4619-9e19-9b2dfd4af40c" containerName="kube-state-metrics" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.674241 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9961e9d-0fa2-4619-9e19-9b2dfd4af40c" containerName="kube-state-metrics" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.674444 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7e7e469-ed0d-463c-839c-95427c632d5b" containerName="dnsmasq-dns" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.674463 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9961e9d-0fa2-4619-9e19-9b2dfd4af40c" containerName="kube-state-metrics" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.675074 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.677161 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.677436 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.694592 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.69457309 podStartE2EDuration="2.69457309s" podCreationTimestamp="2025-12-06 16:05:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:05:24.664052228 +0000 UTC m=+1164.554931804" watchObservedRunningTime="2025-12-06 16:05:24.69457309 +0000 UTC m=+1164.585452666" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.713300 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.763951 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/8d3a2b2d-c016-4948-83c5-f8fcae8678d8-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"8d3a2b2d-c016-4948-83c5-f8fcae8678d8\") " pod="openstack/kube-state-metrics-0" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.764243 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d3a2b2d-c016-4948-83c5-f8fcae8678d8-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"8d3a2b2d-c016-4948-83c5-f8fcae8678d8\") " pod="openstack/kube-state-metrics-0" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.764284 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d3a2b2d-c016-4948-83c5-f8fcae8678d8-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"8d3a2b2d-c016-4948-83c5-f8fcae8678d8\") " pod="openstack/kube-state-metrics-0" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.764326 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kqqw\" (UniqueName: \"kubernetes.io/projected/8d3a2b2d-c016-4948-83c5-f8fcae8678d8-kube-api-access-9kqqw\") pod \"kube-state-metrics-0\" (UID: \"8d3a2b2d-c016-4948-83c5-f8fcae8678d8\") " pod="openstack/kube-state-metrics-0" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.864883 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d3a2b2d-c016-4948-83c5-f8fcae8678d8-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"8d3a2b2d-c016-4948-83c5-f8fcae8678d8\") " pod="openstack/kube-state-metrics-0" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.864935 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d3a2b2d-c016-4948-83c5-f8fcae8678d8-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"8d3a2b2d-c016-4948-83c5-f8fcae8678d8\") " pod="openstack/kube-state-metrics-0" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.864982 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kqqw\" (UniqueName: \"kubernetes.io/projected/8d3a2b2d-c016-4948-83c5-f8fcae8678d8-kube-api-access-9kqqw\") pod \"kube-state-metrics-0\" (UID: \"8d3a2b2d-c016-4948-83c5-f8fcae8678d8\") " pod="openstack/kube-state-metrics-0" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.865064 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/8d3a2b2d-c016-4948-83c5-f8fcae8678d8-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"8d3a2b2d-c016-4948-83c5-f8fcae8678d8\") " pod="openstack/kube-state-metrics-0" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.871828 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/8d3a2b2d-c016-4948-83c5-f8fcae8678d8-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"8d3a2b2d-c016-4948-83c5-f8fcae8678d8\") " pod="openstack/kube-state-metrics-0" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.872031 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d3a2b2d-c016-4948-83c5-f8fcae8678d8-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"8d3a2b2d-c016-4948-83c5-f8fcae8678d8\") " pod="openstack/kube-state-metrics-0" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.875882 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d3a2b2d-c016-4948-83c5-f8fcae8678d8-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"8d3a2b2d-c016-4948-83c5-f8fcae8678d8\") " pod="openstack/kube-state-metrics-0" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.888999 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kqqw\" (UniqueName: \"kubernetes.io/projected/8d3a2b2d-c016-4948-83c5-f8fcae8678d8-kube-api-access-9kqqw\") pod \"kube-state-metrics-0\" (UID: \"8d3a2b2d-c016-4948-83c5-f8fcae8678d8\") " pod="openstack/kube-state-metrics-0" Dec 06 16:05:24 crc kubenswrapper[4813]: I1206 16:05:24.995319 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 16:05:25 crc kubenswrapper[4813]: I1206 16:05:25.535708 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 16:05:25 crc kubenswrapper[4813]: W1206 16:05:25.543982 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d3a2b2d_c016_4948_83c5_f8fcae8678d8.slice/crio-76d2fe5727616317199bacb4d4f7ca5483e8f55b91ae29e13481edc3e0042d69 WatchSource:0}: Error finding container 76d2fe5727616317199bacb4d4f7ca5483e8f55b91ae29e13481edc3e0042d69: Status 404 returned error can't find the container with id 76d2fe5727616317199bacb4d4f7ca5483e8f55b91ae29e13481edc3e0042d69 Dec 06 16:05:25 crc kubenswrapper[4813]: I1206 16:05:25.565036 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8d3a2b2d-c016-4948-83c5-f8fcae8678d8","Type":"ContainerStarted","Data":"76d2fe5727616317199bacb4d4f7ca5483e8f55b91ae29e13481edc3e0042d69"} Dec 06 16:05:25 crc kubenswrapper[4813]: I1206 16:05:25.959534 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xwdwl" Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.001618 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fc8a5f4-00b7-434b-b375-74d5a95e3fd1-combined-ca-bundle\") pod \"8fc8a5f4-00b7-434b-b375-74d5a95e3fd1\" (UID: \"8fc8a5f4-00b7-434b-b375-74d5a95e3fd1\") " Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.001665 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fc8a5f4-00b7-434b-b375-74d5a95e3fd1-config-data\") pod \"8fc8a5f4-00b7-434b-b375-74d5a95e3fd1\" (UID: \"8fc8a5f4-00b7-434b-b375-74d5a95e3fd1\") " Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.001739 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r69zj\" (UniqueName: \"kubernetes.io/projected/8fc8a5f4-00b7-434b-b375-74d5a95e3fd1-kube-api-access-r69zj\") pod \"8fc8a5f4-00b7-434b-b375-74d5a95e3fd1\" (UID: \"8fc8a5f4-00b7-434b-b375-74d5a95e3fd1\") " Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.001846 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fc8a5f4-00b7-434b-b375-74d5a95e3fd1-scripts\") pod \"8fc8a5f4-00b7-434b-b375-74d5a95e3fd1\" (UID: \"8fc8a5f4-00b7-434b-b375-74d5a95e3fd1\") " Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.005707 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fc8a5f4-00b7-434b-b375-74d5a95e3fd1-kube-api-access-r69zj" (OuterVolumeSpecName: "kube-api-access-r69zj") pod "8fc8a5f4-00b7-434b-b375-74d5a95e3fd1" (UID: "8fc8a5f4-00b7-434b-b375-74d5a95e3fd1"). InnerVolumeSpecName "kube-api-access-r69zj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.006687 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fc8a5f4-00b7-434b-b375-74d5a95e3fd1-scripts" (OuterVolumeSpecName: "scripts") pod "8fc8a5f4-00b7-434b-b375-74d5a95e3fd1" (UID: "8fc8a5f4-00b7-434b-b375-74d5a95e3fd1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.045400 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fc8a5f4-00b7-434b-b375-74d5a95e3fd1-config-data" (OuterVolumeSpecName: "config-data") pod "8fc8a5f4-00b7-434b-b375-74d5a95e3fd1" (UID: "8fc8a5f4-00b7-434b-b375-74d5a95e3fd1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.068243 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fc8a5f4-00b7-434b-b375-74d5a95e3fd1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8fc8a5f4-00b7-434b-b375-74d5a95e3fd1" (UID: "8fc8a5f4-00b7-434b-b375-74d5a95e3fd1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.104567 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fc8a5f4-00b7-434b-b375-74d5a95e3fd1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.104608 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fc8a5f4-00b7-434b-b375-74d5a95e3fd1-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.104618 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r69zj\" (UniqueName: \"kubernetes.io/projected/8fc8a5f4-00b7-434b-b375-74d5a95e3fd1-kube-api-access-r69zj\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.104628 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fc8a5f4-00b7-434b-b375-74d5a95e3fd1-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.345473 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.345986 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3860b8fd-910e-472c-af7e-22021b9c5f42" containerName="ceilometer-central-agent" containerID="cri-o://2b1742606d409605cb6948ff573d795ff2b4c04a8d7825e8a117f90e873f7472" gracePeriod=30 Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.346091 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3860b8fd-910e-472c-af7e-22021b9c5f42" containerName="proxy-httpd" containerID="cri-o://258aee34ba8d549328acc52f2ed0f26867e1376d55703490936bf49c0e56ef3a" gracePeriod=30 Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.346126 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3860b8fd-910e-472c-af7e-22021b9c5f42" containerName="sg-core" containerID="cri-o://67331f1b42d28f68de93ac94725527090c4590e29a20c3f9e232a471759d8b2b" gracePeriod=30 Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.346154 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3860b8fd-910e-472c-af7e-22021b9c5f42" containerName="ceilometer-notification-agent" containerID="cri-o://ea14b6dcb07649c817b1db8a8ab1402ac31c0fe6aa8e7e0e9b4866d917a94aa8" gracePeriod=30 Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.498632 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9961e9d-0fa2-4619-9e19-9b2dfd4af40c" path="/var/lib/kubelet/pods/c9961e9d-0fa2-4619-9e19-9b2dfd4af40c/volumes" Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.499146 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7e7e469-ed0d-463c-839c-95427c632d5b" path="/var/lib/kubelet/pods/d7e7e469-ed0d-463c-839c-95427c632d5b/volumes" Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.580739 4813 generic.go:334] "Generic (PLEG): container finished" podID="3860b8fd-910e-472c-af7e-22021b9c5f42" containerID="258aee34ba8d549328acc52f2ed0f26867e1376d55703490936bf49c0e56ef3a" exitCode=0 Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.580769 4813 generic.go:334] "Generic (PLEG): container finished" podID="3860b8fd-910e-472c-af7e-22021b9c5f42" containerID="67331f1b42d28f68de93ac94725527090c4590e29a20c3f9e232a471759d8b2b" exitCode=2 Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.580779 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3860b8fd-910e-472c-af7e-22021b9c5f42","Type":"ContainerDied","Data":"258aee34ba8d549328acc52f2ed0f26867e1376d55703490936bf49c0e56ef3a"} Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.580847 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3860b8fd-910e-472c-af7e-22021b9c5f42","Type":"ContainerDied","Data":"67331f1b42d28f68de93ac94725527090c4590e29a20c3f9e232a471759d8b2b"} Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.583394 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xwdwl" Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.583393 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xwdwl" event={"ID":"8fc8a5f4-00b7-434b-b375-74d5a95e3fd1","Type":"ContainerDied","Data":"a9bddc90671279b3d03f13803796c644348acd9afb87e0bee54b0f0685425a07"} Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.583514 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9bddc90671279b3d03f13803796c644348acd9afb87e0bee54b0f0685425a07" Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.585757 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8d3a2b2d-c016-4948-83c5-f8fcae8678d8","Type":"ContainerStarted","Data":"2762523f96d95e872ebb76dfb75ac96893e3f2547a59f0588bf24204a2d8cf79"} Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.586653 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.617929 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.218065332 podStartE2EDuration="2.617909796s" podCreationTimestamp="2025-12-06 16:05:24 +0000 UTC" firstStartedPulling="2025-12-06 16:05:25.546583162 +0000 UTC m=+1165.437462738" lastFinishedPulling="2025-12-06 16:05:25.946427626 +0000 UTC m=+1165.837307202" observedRunningTime="2025-12-06 16:05:26.606968108 +0000 UTC m=+1166.497847694" watchObservedRunningTime="2025-12-06 16:05:26.617909796 +0000 UTC m=+1166.508789372" Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.839413 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.839655 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="97c18b03-4126-4af3-a790-b7450e54c418" containerName="nova-api-log" containerID="cri-o://1fb6e6ef65034b4e720bedb2ef2b57091bf7d57b3931fb0bac81fbfdf396cb1c" gracePeriod=30 Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.840037 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="97c18b03-4126-4af3-a790-b7450e54c418" containerName="nova-api-api" containerID="cri-o://56a4caccafc3b2b9f412eb23ce642e5c300d84069300ba7fa0b382fefe96ef39" gracePeriod=30 Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.844278 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.844448 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="7da70874-b4ec-4a5a-9061-af3983fd7b88" containerName="nova-scheduler-scheduler" containerID="cri-o://2b1588c231aa5871f88c2b3c13a462416c153666299d3a05e6367d79b9e062ef" gracePeriod=30 Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.896656 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.896857 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f70a63c7-a091-4b73-b86e-805e3529a69b" containerName="nova-metadata-log" containerID="cri-o://e429a1d4e041b9aa8abe15788220b63145b20aff70484a7775f4259b265e456a" gracePeriod=30 Dec 06 16:05:26 crc kubenswrapper[4813]: I1206 16:05:26.897217 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f70a63c7-a091-4b73-b86e-805e3529a69b" containerName="nova-metadata-metadata" containerID="cri-o://0d54482c9adab1a07bb808250d0b1b6f7d31ce1170140ee8609b9f217dd7333e" gracePeriod=30 Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.408738 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.524903 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttlm8\" (UniqueName: \"kubernetes.io/projected/f70a63c7-a091-4b73-b86e-805e3529a69b-kube-api-access-ttlm8\") pod \"f70a63c7-a091-4b73-b86e-805e3529a69b\" (UID: \"f70a63c7-a091-4b73-b86e-805e3529a69b\") " Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.524950 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f70a63c7-a091-4b73-b86e-805e3529a69b-nova-metadata-tls-certs\") pod \"f70a63c7-a091-4b73-b86e-805e3529a69b\" (UID: \"f70a63c7-a091-4b73-b86e-805e3529a69b\") " Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.524997 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f70a63c7-a091-4b73-b86e-805e3529a69b-logs\") pod \"f70a63c7-a091-4b73-b86e-805e3529a69b\" (UID: \"f70a63c7-a091-4b73-b86e-805e3529a69b\") " Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.525035 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f70a63c7-a091-4b73-b86e-805e3529a69b-combined-ca-bundle\") pod \"f70a63c7-a091-4b73-b86e-805e3529a69b\" (UID: \"f70a63c7-a091-4b73-b86e-805e3529a69b\") " Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.525077 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f70a63c7-a091-4b73-b86e-805e3529a69b-config-data\") pod \"f70a63c7-a091-4b73-b86e-805e3529a69b\" (UID: \"f70a63c7-a091-4b73-b86e-805e3529a69b\") " Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.525765 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f70a63c7-a091-4b73-b86e-805e3529a69b-logs" (OuterVolumeSpecName: "logs") pod "f70a63c7-a091-4b73-b86e-805e3529a69b" (UID: "f70a63c7-a091-4b73-b86e-805e3529a69b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.530475 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f70a63c7-a091-4b73-b86e-805e3529a69b-kube-api-access-ttlm8" (OuterVolumeSpecName: "kube-api-access-ttlm8") pod "f70a63c7-a091-4b73-b86e-805e3529a69b" (UID: "f70a63c7-a091-4b73-b86e-805e3529a69b"). InnerVolumeSpecName "kube-api-access-ttlm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.558939 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f70a63c7-a091-4b73-b86e-805e3529a69b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f70a63c7-a091-4b73-b86e-805e3529a69b" (UID: "f70a63c7-a091-4b73-b86e-805e3529a69b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.559321 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f70a63c7-a091-4b73-b86e-805e3529a69b-config-data" (OuterVolumeSpecName: "config-data") pod "f70a63c7-a091-4b73-b86e-805e3529a69b" (UID: "f70a63c7-a091-4b73-b86e-805e3529a69b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.581463 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f70a63c7-a091-4b73-b86e-805e3529a69b-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "f70a63c7-a091-4b73-b86e-805e3529a69b" (UID: "f70a63c7-a091-4b73-b86e-805e3529a69b"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.621367 4813 generic.go:334] "Generic (PLEG): container finished" podID="3860b8fd-910e-472c-af7e-22021b9c5f42" containerID="2b1742606d409605cb6948ff573d795ff2b4c04a8d7825e8a117f90e873f7472" exitCode=0 Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.621449 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3860b8fd-910e-472c-af7e-22021b9c5f42","Type":"ContainerDied","Data":"2b1742606d409605cb6948ff573d795ff2b4c04a8d7825e8a117f90e873f7472"} Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.624901 4813 generic.go:334] "Generic (PLEG): container finished" podID="f70a63c7-a091-4b73-b86e-805e3529a69b" containerID="0d54482c9adab1a07bb808250d0b1b6f7d31ce1170140ee8609b9f217dd7333e" exitCode=0 Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.625026 4813 generic.go:334] "Generic (PLEG): container finished" podID="f70a63c7-a091-4b73-b86e-805e3529a69b" containerID="e429a1d4e041b9aa8abe15788220b63145b20aff70484a7775f4259b265e456a" exitCode=143 Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.624959 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.624974 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f70a63c7-a091-4b73-b86e-805e3529a69b","Type":"ContainerDied","Data":"0d54482c9adab1a07bb808250d0b1b6f7d31ce1170140ee8609b9f217dd7333e"} Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.625137 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f70a63c7-a091-4b73-b86e-805e3529a69b","Type":"ContainerDied","Data":"e429a1d4e041b9aa8abe15788220b63145b20aff70484a7775f4259b265e456a"} Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.625153 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f70a63c7-a091-4b73-b86e-805e3529a69b","Type":"ContainerDied","Data":"54de8c8e2d8e340e24aff8768c4dbf080cc4293d1fb4b0502bf322a5b7b142ed"} Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.625172 4813 scope.go:117] "RemoveContainer" containerID="0d54482c9adab1a07bb808250d0b1b6f7d31ce1170140ee8609b9f217dd7333e" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.626753 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f70a63c7-a091-4b73-b86e-805e3529a69b-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.626840 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttlm8\" (UniqueName: \"kubernetes.io/projected/f70a63c7-a091-4b73-b86e-805e3529a69b-kube-api-access-ttlm8\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.626897 4813 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f70a63c7-a091-4b73-b86e-805e3529a69b-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.626969 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f70a63c7-a091-4b73-b86e-805e3529a69b-logs\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.627024 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f70a63c7-a091-4b73-b86e-805e3529a69b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.639475 4813 generic.go:334] "Generic (PLEG): container finished" podID="97c18b03-4126-4af3-a790-b7450e54c418" containerID="1fb6e6ef65034b4e720bedb2ef2b57091bf7d57b3931fb0bac81fbfdf396cb1c" exitCode=143 Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.639793 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"97c18b03-4126-4af3-a790-b7450e54c418","Type":"ContainerDied","Data":"1fb6e6ef65034b4e720bedb2ef2b57091bf7d57b3931fb0bac81fbfdf396cb1c"} Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.681837 4813 scope.go:117] "RemoveContainer" containerID="e429a1d4e041b9aa8abe15788220b63145b20aff70484a7775f4259b265e456a" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.701320 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.722484 4813 scope.go:117] "RemoveContainer" containerID="0d54482c9adab1a07bb808250d0b1b6f7d31ce1170140ee8609b9f217dd7333e" Dec 06 16:05:27 crc kubenswrapper[4813]: E1206 16:05:27.729657 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d54482c9adab1a07bb808250d0b1b6f7d31ce1170140ee8609b9f217dd7333e\": container with ID starting with 0d54482c9adab1a07bb808250d0b1b6f7d31ce1170140ee8609b9f217dd7333e not found: ID does not exist" containerID="0d54482c9adab1a07bb808250d0b1b6f7d31ce1170140ee8609b9f217dd7333e" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.729699 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d54482c9adab1a07bb808250d0b1b6f7d31ce1170140ee8609b9f217dd7333e"} err="failed to get container status \"0d54482c9adab1a07bb808250d0b1b6f7d31ce1170140ee8609b9f217dd7333e\": rpc error: code = NotFound desc = could not find container \"0d54482c9adab1a07bb808250d0b1b6f7d31ce1170140ee8609b9f217dd7333e\": container with ID starting with 0d54482c9adab1a07bb808250d0b1b6f7d31ce1170140ee8609b9f217dd7333e not found: ID does not exist" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.729723 4813 scope.go:117] "RemoveContainer" containerID="e429a1d4e041b9aa8abe15788220b63145b20aff70484a7775f4259b265e456a" Dec 06 16:05:27 crc kubenswrapper[4813]: E1206 16:05:27.730417 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e429a1d4e041b9aa8abe15788220b63145b20aff70484a7775f4259b265e456a\": container with ID starting with e429a1d4e041b9aa8abe15788220b63145b20aff70484a7775f4259b265e456a not found: ID does not exist" containerID="e429a1d4e041b9aa8abe15788220b63145b20aff70484a7775f4259b265e456a" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.730436 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e429a1d4e041b9aa8abe15788220b63145b20aff70484a7775f4259b265e456a"} err="failed to get container status \"e429a1d4e041b9aa8abe15788220b63145b20aff70484a7775f4259b265e456a\": rpc error: code = NotFound desc = could not find container \"e429a1d4e041b9aa8abe15788220b63145b20aff70484a7775f4259b265e456a\": container with ID starting with e429a1d4e041b9aa8abe15788220b63145b20aff70484a7775f4259b265e456a not found: ID does not exist" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.730449 4813 scope.go:117] "RemoveContainer" containerID="0d54482c9adab1a07bb808250d0b1b6f7d31ce1170140ee8609b9f217dd7333e" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.735589 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d54482c9adab1a07bb808250d0b1b6f7d31ce1170140ee8609b9f217dd7333e"} err="failed to get container status \"0d54482c9adab1a07bb808250d0b1b6f7d31ce1170140ee8609b9f217dd7333e\": rpc error: code = NotFound desc = could not find container \"0d54482c9adab1a07bb808250d0b1b6f7d31ce1170140ee8609b9f217dd7333e\": container with ID starting with 0d54482c9adab1a07bb808250d0b1b6f7d31ce1170140ee8609b9f217dd7333e not found: ID does not exist" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.735626 4813 scope.go:117] "RemoveContainer" containerID="e429a1d4e041b9aa8abe15788220b63145b20aff70484a7775f4259b265e456a" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.738620 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.739361 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e429a1d4e041b9aa8abe15788220b63145b20aff70484a7775f4259b265e456a"} err="failed to get container status \"e429a1d4e041b9aa8abe15788220b63145b20aff70484a7775f4259b265e456a\": rpc error: code = NotFound desc = could not find container \"e429a1d4e041b9aa8abe15788220b63145b20aff70484a7775f4259b265e456a\": container with ID starting with e429a1d4e041b9aa8abe15788220b63145b20aff70484a7775f4259b265e456a not found: ID does not exist" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.752314 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 06 16:05:27 crc kubenswrapper[4813]: E1206 16:05:27.752675 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f70a63c7-a091-4b73-b86e-805e3529a69b" containerName="nova-metadata-log" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.752692 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f70a63c7-a091-4b73-b86e-805e3529a69b" containerName="nova-metadata-log" Dec 06 16:05:27 crc kubenswrapper[4813]: E1206 16:05:27.752705 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f70a63c7-a091-4b73-b86e-805e3529a69b" containerName="nova-metadata-metadata" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.752713 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f70a63c7-a091-4b73-b86e-805e3529a69b" containerName="nova-metadata-metadata" Dec 06 16:05:27 crc kubenswrapper[4813]: E1206 16:05:27.752724 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fc8a5f4-00b7-434b-b375-74d5a95e3fd1" containerName="nova-manage" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.752730 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fc8a5f4-00b7-434b-b375-74d5a95e3fd1" containerName="nova-manage" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.752892 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f70a63c7-a091-4b73-b86e-805e3529a69b" containerName="nova-metadata-log" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.752912 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f70a63c7-a091-4b73-b86e-805e3529a69b" containerName="nova-metadata-metadata" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.752936 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fc8a5f4-00b7-434b-b375-74d5a95e3fd1" containerName="nova-manage" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.761672 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.780766 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.781649 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.785671 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.931855 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99-logs\") pod \"nova-metadata-0\" (UID: \"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99\") " pod="openstack/nova-metadata-0" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.931931 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfqfk\" (UniqueName: \"kubernetes.io/projected/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99-kube-api-access-jfqfk\") pod \"nova-metadata-0\" (UID: \"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99\") " pod="openstack/nova-metadata-0" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.931979 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99-config-data\") pod \"nova-metadata-0\" (UID: \"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99\") " pod="openstack/nova-metadata-0" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.931998 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99\") " pod="openstack/nova-metadata-0" Dec 06 16:05:27 crc kubenswrapper[4813]: I1206 16:05:27.932028 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99\") " pod="openstack/nova-metadata-0" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.033055 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfqfk\" (UniqueName: \"kubernetes.io/projected/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99-kube-api-access-jfqfk\") pod \"nova-metadata-0\" (UID: \"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99\") " pod="openstack/nova-metadata-0" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.033133 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99-config-data\") pod \"nova-metadata-0\" (UID: \"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99\") " pod="openstack/nova-metadata-0" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.033152 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99\") " pod="openstack/nova-metadata-0" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.033181 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99\") " pod="openstack/nova-metadata-0" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.033274 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99-logs\") pod \"nova-metadata-0\" (UID: \"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99\") " pod="openstack/nova-metadata-0" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.034035 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99-logs\") pod \"nova-metadata-0\" (UID: \"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99\") " pod="openstack/nova-metadata-0" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.040490 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99\") " pod="openstack/nova-metadata-0" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.048778 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99\") " pod="openstack/nova-metadata-0" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.066742 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99-config-data\") pod \"nova-metadata-0\" (UID: \"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99\") " pod="openstack/nova-metadata-0" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.079777 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfqfk\" (UniqueName: \"kubernetes.io/projected/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99-kube-api-access-jfqfk\") pod \"nova-metadata-0\" (UID: \"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99\") " pod="openstack/nova-metadata-0" Dec 06 16:05:28 crc kubenswrapper[4813]: E1206 16:05:28.102307 4813 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2b1588c231aa5871f88c2b3c13a462416c153666299d3a05e6367d79b9e062ef is running failed: container process not found" containerID="2b1588c231aa5871f88c2b3c13a462416c153666299d3a05e6367d79b9e062ef" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 16:05:28 crc kubenswrapper[4813]: E1206 16:05:28.103720 4813 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2b1588c231aa5871f88c2b3c13a462416c153666299d3a05e6367d79b9e062ef is running failed: container process not found" containerID="2b1588c231aa5871f88c2b3c13a462416c153666299d3a05e6367d79b9e062ef" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.103764 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 16:05:28 crc kubenswrapper[4813]: E1206 16:05:28.104027 4813 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2b1588c231aa5871f88c2b3c13a462416c153666299d3a05e6367d79b9e062ef is running failed: container process not found" containerID="2b1588c231aa5871f88c2b3c13a462416c153666299d3a05e6367d79b9e062ef" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 16:05:28 crc kubenswrapper[4813]: E1206 16:05:28.104054 4813 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2b1588c231aa5871f88c2b3c13a462416c153666299d3a05e6367d79b9e062ef is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="7da70874-b4ec-4a5a-9061-af3983fd7b88" containerName="nova-scheduler-scheduler" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.437838 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 16:05:28 crc kubenswrapper[4813]: W1206 16:05:28.483300 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff13cc9e_77d0_46a8_9f98_98dd3a75cc99.slice/crio-74933231fba764bb2aaf2412fb2ff42d397750898954c85ce628debcd914ec63 WatchSource:0}: Error finding container 74933231fba764bb2aaf2412fb2ff42d397750898954c85ce628debcd914ec63: Status 404 returned error can't find the container with id 74933231fba764bb2aaf2412fb2ff42d397750898954c85ce628debcd914ec63 Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.502462 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f70a63c7-a091-4b73-b86e-805e3529a69b" path="/var/lib/kubelet/pods/f70a63c7-a091-4b73-b86e-805e3529a69b/volumes" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.507385 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.543755 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7da70874-b4ec-4a5a-9061-af3983fd7b88-combined-ca-bundle\") pod \"7da70874-b4ec-4a5a-9061-af3983fd7b88\" (UID: \"7da70874-b4ec-4a5a-9061-af3983fd7b88\") " Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.543859 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxnx9\" (UniqueName: \"kubernetes.io/projected/7da70874-b4ec-4a5a-9061-af3983fd7b88-kube-api-access-rxnx9\") pod \"7da70874-b4ec-4a5a-9061-af3983fd7b88\" (UID: \"7da70874-b4ec-4a5a-9061-af3983fd7b88\") " Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.543876 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7da70874-b4ec-4a5a-9061-af3983fd7b88-config-data\") pod \"7da70874-b4ec-4a5a-9061-af3983fd7b88\" (UID: \"7da70874-b4ec-4a5a-9061-af3983fd7b88\") " Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.550850 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7da70874-b4ec-4a5a-9061-af3983fd7b88-kube-api-access-rxnx9" (OuterVolumeSpecName: "kube-api-access-rxnx9") pod "7da70874-b4ec-4a5a-9061-af3983fd7b88" (UID: "7da70874-b4ec-4a5a-9061-af3983fd7b88"). InnerVolumeSpecName "kube-api-access-rxnx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.583681 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7da70874-b4ec-4a5a-9061-af3983fd7b88-config-data" (OuterVolumeSpecName: "config-data") pod "7da70874-b4ec-4a5a-9061-af3983fd7b88" (UID: "7da70874-b4ec-4a5a-9061-af3983fd7b88"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.584943 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7da70874-b4ec-4a5a-9061-af3983fd7b88-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7da70874-b4ec-4a5a-9061-af3983fd7b88" (UID: "7da70874-b4ec-4a5a-9061-af3983fd7b88"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.646129 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxnx9\" (UniqueName: \"kubernetes.io/projected/7da70874-b4ec-4a5a-9061-af3983fd7b88-kube-api-access-rxnx9\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.646160 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7da70874-b4ec-4a5a-9061-af3983fd7b88-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.646170 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7da70874-b4ec-4a5a-9061-af3983fd7b88-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.657921 4813 generic.go:334] "Generic (PLEG): container finished" podID="7da70874-b4ec-4a5a-9061-af3983fd7b88" containerID="2b1588c231aa5871f88c2b3c13a462416c153666299d3a05e6367d79b9e062ef" exitCode=0 Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.657980 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7da70874-b4ec-4a5a-9061-af3983fd7b88","Type":"ContainerDied","Data":"2b1588c231aa5871f88c2b3c13a462416c153666299d3a05e6367d79b9e062ef"} Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.658006 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7da70874-b4ec-4a5a-9061-af3983fd7b88","Type":"ContainerDied","Data":"7ad4b9b48589e67f08bd44eda3837a9134e4d428addaf896ccb48425517bbc69"} Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.658025 4813 scope.go:117] "RemoveContainer" containerID="2b1588c231aa5871f88c2b3c13a462416c153666299d3a05e6367d79b9e062ef" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.658120 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.661363 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99","Type":"ContainerStarted","Data":"6bb20d0d07b325fc233067fe52f03c18e26f7673230c3edb8903cb9501d81c40"} Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.661388 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99","Type":"ContainerStarted","Data":"74933231fba764bb2aaf2412fb2ff42d397750898954c85ce628debcd914ec63"} Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.721343 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.724924 4813 scope.go:117] "RemoveContainer" containerID="2b1588c231aa5871f88c2b3c13a462416c153666299d3a05e6367d79b9e062ef" Dec 06 16:05:28 crc kubenswrapper[4813]: E1206 16:05:28.725391 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b1588c231aa5871f88c2b3c13a462416c153666299d3a05e6367d79b9e062ef\": container with ID starting with 2b1588c231aa5871f88c2b3c13a462416c153666299d3a05e6367d79b9e062ef not found: ID does not exist" containerID="2b1588c231aa5871f88c2b3c13a462416c153666299d3a05e6367d79b9e062ef" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.725416 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b1588c231aa5871f88c2b3c13a462416c153666299d3a05e6367d79b9e062ef"} err="failed to get container status \"2b1588c231aa5871f88c2b3c13a462416c153666299d3a05e6367d79b9e062ef\": rpc error: code = NotFound desc = could not find container \"2b1588c231aa5871f88c2b3c13a462416c153666299d3a05e6367d79b9e062ef\": container with ID starting with 2b1588c231aa5871f88c2b3c13a462416c153666299d3a05e6367d79b9e062ef not found: ID does not exist" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.729335 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.736116 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 16:05:28 crc kubenswrapper[4813]: E1206 16:05:28.736494 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7da70874-b4ec-4a5a-9061-af3983fd7b88" containerName="nova-scheduler-scheduler" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.736511 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="7da70874-b4ec-4a5a-9061-af3983fd7b88" containerName="nova-scheduler-scheduler" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.736670 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="7da70874-b4ec-4a5a-9061-af3983fd7b88" containerName="nova-scheduler-scheduler" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.737366 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.742120 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.749935 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.849603 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5clb\" (UniqueName: \"kubernetes.io/projected/4af45f0b-4427-44d1-a612-ece893ad9b55-kube-api-access-z5clb\") pod \"nova-scheduler-0\" (UID: \"4af45f0b-4427-44d1-a612-ece893ad9b55\") " pod="openstack/nova-scheduler-0" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.849657 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af45f0b-4427-44d1-a612-ece893ad9b55-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4af45f0b-4427-44d1-a612-ece893ad9b55\") " pod="openstack/nova-scheduler-0" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.849734 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4af45f0b-4427-44d1-a612-ece893ad9b55-config-data\") pod \"nova-scheduler-0\" (UID: \"4af45f0b-4427-44d1-a612-ece893ad9b55\") " pod="openstack/nova-scheduler-0" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.951159 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5clb\" (UniqueName: \"kubernetes.io/projected/4af45f0b-4427-44d1-a612-ece893ad9b55-kube-api-access-z5clb\") pod \"nova-scheduler-0\" (UID: \"4af45f0b-4427-44d1-a612-ece893ad9b55\") " pod="openstack/nova-scheduler-0" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.951210 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af45f0b-4427-44d1-a612-ece893ad9b55-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4af45f0b-4427-44d1-a612-ece893ad9b55\") " pod="openstack/nova-scheduler-0" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.951274 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4af45f0b-4427-44d1-a612-ece893ad9b55-config-data\") pod \"nova-scheduler-0\" (UID: \"4af45f0b-4427-44d1-a612-ece893ad9b55\") " pod="openstack/nova-scheduler-0" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.956024 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4af45f0b-4427-44d1-a612-ece893ad9b55-config-data\") pod \"nova-scheduler-0\" (UID: \"4af45f0b-4427-44d1-a612-ece893ad9b55\") " pod="openstack/nova-scheduler-0" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.956539 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af45f0b-4427-44d1-a612-ece893ad9b55-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4af45f0b-4427-44d1-a612-ece893ad9b55\") " pod="openstack/nova-scheduler-0" Dec 06 16:05:28 crc kubenswrapper[4813]: I1206 16:05:28.970533 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5clb\" (UniqueName: \"kubernetes.io/projected/4af45f0b-4427-44d1-a612-ece893ad9b55-kube-api-access-z5clb\") pod \"nova-scheduler-0\" (UID: \"4af45f0b-4427-44d1-a612-ece893ad9b55\") " pod="openstack/nova-scheduler-0" Dec 06 16:05:29 crc kubenswrapper[4813]: I1206 16:05:29.126600 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 16:05:29 crc kubenswrapper[4813]: I1206 16:05:29.621347 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 16:05:29 crc kubenswrapper[4813]: W1206 16:05:29.624552 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4af45f0b_4427_44d1_a612_ece893ad9b55.slice/crio-9f3e2b1953cea6dcacc64a91ace828be60e134a35d4e91b08fac0dde15ae985d WatchSource:0}: Error finding container 9f3e2b1953cea6dcacc64a91ace828be60e134a35d4e91b08fac0dde15ae985d: Status 404 returned error can't find the container with id 9f3e2b1953cea6dcacc64a91ace828be60e134a35d4e91b08fac0dde15ae985d Dec 06 16:05:29 crc kubenswrapper[4813]: I1206 16:05:29.674498 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99","Type":"ContainerStarted","Data":"7facf55ea2214f87399157647a3498e61a5829cf2d0cad852b6563da23382d84"} Dec 06 16:05:29 crc kubenswrapper[4813]: I1206 16:05:29.676733 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4af45f0b-4427-44d1-a612-ece893ad9b55","Type":"ContainerStarted","Data":"9f3e2b1953cea6dcacc64a91ace828be60e134a35d4e91b08fac0dde15ae985d"} Dec 06 16:05:29 crc kubenswrapper[4813]: I1206 16:05:29.699021 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.698999874 podStartE2EDuration="2.698999874s" podCreationTimestamp="2025-12-06 16:05:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:05:29.694932426 +0000 UTC m=+1169.585812012" watchObservedRunningTime="2025-12-06 16:05:29.698999874 +0000 UTC m=+1169.589879450" Dec 06 16:05:30 crc kubenswrapper[4813]: I1206 16:05:30.522189 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7da70874-b4ec-4a5a-9061-af3983fd7b88" path="/var/lib/kubelet/pods/7da70874-b4ec-4a5a-9061-af3983fd7b88/volumes" Dec 06 16:05:30 crc kubenswrapper[4813]: I1206 16:05:30.684897 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4af45f0b-4427-44d1-a612-ece893ad9b55","Type":"ContainerStarted","Data":"c2354e660285dee3f5e8b0c70d6303463fc1167d6c465cf4a40d704587343afc"} Dec 06 16:05:30 crc kubenswrapper[4813]: I1206 16:05:30.685946 4813 generic.go:334] "Generic (PLEG): container finished" podID="4d9f1a15-0cf8-45e2-b4f4-d15d257d7009" containerID="bd8568faea71fd736f6a65c700cae646473782de03144171c2d8ade2038b5179" exitCode=0 Dec 06 16:05:30 crc kubenswrapper[4813]: I1206 16:05:30.686076 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-gv4hp" event={"ID":"4d9f1a15-0cf8-45e2-b4f4-d15d257d7009","Type":"ContainerDied","Data":"bd8568faea71fd736f6a65c700cae646473782de03144171c2d8ade2038b5179"} Dec 06 16:05:30 crc kubenswrapper[4813]: I1206 16:05:30.713367 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.713352046 podStartE2EDuration="2.713352046s" podCreationTimestamp="2025-12-06 16:05:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:05:30.706544806 +0000 UTC m=+1170.597424382" watchObservedRunningTime="2025-12-06 16:05:30.713352046 +0000 UTC m=+1170.604231622" Dec 06 16:05:31 crc kubenswrapper[4813]: I1206 16:05:31.702563 4813 generic.go:334] "Generic (PLEG): container finished" podID="97c18b03-4126-4af3-a790-b7450e54c418" containerID="56a4caccafc3b2b9f412eb23ce642e5c300d84069300ba7fa0b382fefe96ef39" exitCode=0 Dec 06 16:05:31 crc kubenswrapper[4813]: I1206 16:05:31.702824 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"97c18b03-4126-4af3-a790-b7450e54c418","Type":"ContainerDied","Data":"56a4caccafc3b2b9f412eb23ce642e5c300d84069300ba7fa0b382fefe96ef39"} Dec 06 16:05:31 crc kubenswrapper[4813]: I1206 16:05:31.713471 4813 generic.go:334] "Generic (PLEG): container finished" podID="3860b8fd-910e-472c-af7e-22021b9c5f42" containerID="ea14b6dcb07649c817b1db8a8ab1402ac31c0fe6aa8e7e0e9b4866d917a94aa8" exitCode=0 Dec 06 16:05:31 crc kubenswrapper[4813]: I1206 16:05:31.714299 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3860b8fd-910e-472c-af7e-22021b9c5f42","Type":"ContainerDied","Data":"ea14b6dcb07649c817b1db8a8ab1402ac31c0fe6aa8e7e0e9b4866d917a94aa8"} Dec 06 16:05:31 crc kubenswrapper[4813]: I1206 16:05:31.919800 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 16:05:31 crc kubenswrapper[4813]: I1206 16:05:31.923722 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.012355 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97c18b03-4126-4af3-a790-b7450e54c418-logs\") pod \"97c18b03-4126-4af3-a790-b7450e54c418\" (UID: \"97c18b03-4126-4af3-a790-b7450e54c418\") " Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.012708 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3860b8fd-910e-472c-af7e-22021b9c5f42-log-httpd\") pod \"3860b8fd-910e-472c-af7e-22021b9c5f42\" (UID: \"3860b8fd-910e-472c-af7e-22021b9c5f42\") " Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.012761 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3860b8fd-910e-472c-af7e-22021b9c5f42-sg-core-conf-yaml\") pod \"3860b8fd-910e-472c-af7e-22021b9c5f42\" (UID: \"3860b8fd-910e-472c-af7e-22021b9c5f42\") " Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.012797 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nvf4\" (UniqueName: \"kubernetes.io/projected/97c18b03-4126-4af3-a790-b7450e54c418-kube-api-access-8nvf4\") pod \"97c18b03-4126-4af3-a790-b7450e54c418\" (UID: \"97c18b03-4126-4af3-a790-b7450e54c418\") " Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.012834 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3860b8fd-910e-472c-af7e-22021b9c5f42-config-data\") pod \"3860b8fd-910e-472c-af7e-22021b9c5f42\" (UID: \"3860b8fd-910e-472c-af7e-22021b9c5f42\") " Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.012862 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sd55v\" (UniqueName: \"kubernetes.io/projected/3860b8fd-910e-472c-af7e-22021b9c5f42-kube-api-access-sd55v\") pod \"3860b8fd-910e-472c-af7e-22021b9c5f42\" (UID: \"3860b8fd-910e-472c-af7e-22021b9c5f42\") " Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.012869 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97c18b03-4126-4af3-a790-b7450e54c418-logs" (OuterVolumeSpecName: "logs") pod "97c18b03-4126-4af3-a790-b7450e54c418" (UID: "97c18b03-4126-4af3-a790-b7450e54c418"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.012904 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3860b8fd-910e-472c-af7e-22021b9c5f42-scripts\") pod \"3860b8fd-910e-472c-af7e-22021b9c5f42\" (UID: \"3860b8fd-910e-472c-af7e-22021b9c5f42\") " Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.012972 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3860b8fd-910e-472c-af7e-22021b9c5f42-combined-ca-bundle\") pod \"3860b8fd-910e-472c-af7e-22021b9c5f42\" (UID: \"3860b8fd-910e-472c-af7e-22021b9c5f42\") " Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.012997 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3860b8fd-910e-472c-af7e-22021b9c5f42-run-httpd\") pod \"3860b8fd-910e-472c-af7e-22021b9c5f42\" (UID: \"3860b8fd-910e-472c-af7e-22021b9c5f42\") " Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.013069 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97c18b03-4126-4af3-a790-b7450e54c418-config-data\") pod \"97c18b03-4126-4af3-a790-b7450e54c418\" (UID: \"97c18b03-4126-4af3-a790-b7450e54c418\") " Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.013127 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97c18b03-4126-4af3-a790-b7450e54c418-combined-ca-bundle\") pod \"97c18b03-4126-4af3-a790-b7450e54c418\" (UID: \"97c18b03-4126-4af3-a790-b7450e54c418\") " Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.013593 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97c18b03-4126-4af3-a790-b7450e54c418-logs\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.013702 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3860b8fd-910e-472c-af7e-22021b9c5f42-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3860b8fd-910e-472c-af7e-22021b9c5f42" (UID: "3860b8fd-910e-472c-af7e-22021b9c5f42"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.028968 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3860b8fd-910e-472c-af7e-22021b9c5f42-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3860b8fd-910e-472c-af7e-22021b9c5f42" (UID: "3860b8fd-910e-472c-af7e-22021b9c5f42"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.039720 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97c18b03-4126-4af3-a790-b7450e54c418-kube-api-access-8nvf4" (OuterVolumeSpecName: "kube-api-access-8nvf4") pod "97c18b03-4126-4af3-a790-b7450e54c418" (UID: "97c18b03-4126-4af3-a790-b7450e54c418"). InnerVolumeSpecName "kube-api-access-8nvf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.047492 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3860b8fd-910e-472c-af7e-22021b9c5f42-scripts" (OuterVolumeSpecName: "scripts") pod "3860b8fd-910e-472c-af7e-22021b9c5f42" (UID: "3860b8fd-910e-472c-af7e-22021b9c5f42"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.047571 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3860b8fd-910e-472c-af7e-22021b9c5f42-kube-api-access-sd55v" (OuterVolumeSpecName: "kube-api-access-sd55v") pod "3860b8fd-910e-472c-af7e-22021b9c5f42" (UID: "3860b8fd-910e-472c-af7e-22021b9c5f42"). InnerVolumeSpecName "kube-api-access-sd55v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.074191 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3860b8fd-910e-472c-af7e-22021b9c5f42-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3860b8fd-910e-472c-af7e-22021b9c5f42" (UID: "3860b8fd-910e-472c-af7e-22021b9c5f42"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.077207 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97c18b03-4126-4af3-a790-b7450e54c418-config-data" (OuterVolumeSpecName: "config-data") pod "97c18b03-4126-4af3-a790-b7450e54c418" (UID: "97c18b03-4126-4af3-a790-b7450e54c418"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.077988 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97c18b03-4126-4af3-a790-b7450e54c418-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "97c18b03-4126-4af3-a790-b7450e54c418" (UID: "97c18b03-4126-4af3-a790-b7450e54c418"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.105189 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-gv4hp" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.125887 4813 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3860b8fd-910e-472c-af7e-22021b9c5f42-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.130527 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97c18b03-4126-4af3-a790-b7450e54c418-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.130553 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97c18b03-4126-4af3-a790-b7450e54c418-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.130563 4813 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3860b8fd-910e-472c-af7e-22021b9c5f42-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.130571 4813 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3860b8fd-910e-472c-af7e-22021b9c5f42-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.130582 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nvf4\" (UniqueName: \"kubernetes.io/projected/97c18b03-4126-4af3-a790-b7450e54c418-kube-api-access-8nvf4\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.130592 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sd55v\" (UniqueName: \"kubernetes.io/projected/3860b8fd-910e-472c-af7e-22021b9c5f42-kube-api-access-sd55v\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.130600 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3860b8fd-910e-472c-af7e-22021b9c5f42-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.170331 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3860b8fd-910e-472c-af7e-22021b9c5f42-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3860b8fd-910e-472c-af7e-22021b9c5f42" (UID: "3860b8fd-910e-472c-af7e-22021b9c5f42"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.198939 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3860b8fd-910e-472c-af7e-22021b9c5f42-config-data" (OuterVolumeSpecName: "config-data") pod "3860b8fd-910e-472c-af7e-22021b9c5f42" (UID: "3860b8fd-910e-472c-af7e-22021b9c5f42"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.231276 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d9f1a15-0cf8-45e2-b4f4-d15d257d7009-config-data\") pod \"4d9f1a15-0cf8-45e2-b4f4-d15d257d7009\" (UID: \"4d9f1a15-0cf8-45e2-b4f4-d15d257d7009\") " Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.231466 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d9f1a15-0cf8-45e2-b4f4-d15d257d7009-scripts\") pod \"4d9f1a15-0cf8-45e2-b4f4-d15d257d7009\" (UID: \"4d9f1a15-0cf8-45e2-b4f4-d15d257d7009\") " Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.231496 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d9f1a15-0cf8-45e2-b4f4-d15d257d7009-combined-ca-bundle\") pod \"4d9f1a15-0cf8-45e2-b4f4-d15d257d7009\" (UID: \"4d9f1a15-0cf8-45e2-b4f4-d15d257d7009\") " Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.231542 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4nvt\" (UniqueName: \"kubernetes.io/projected/4d9f1a15-0cf8-45e2-b4f4-d15d257d7009-kube-api-access-d4nvt\") pod \"4d9f1a15-0cf8-45e2-b4f4-d15d257d7009\" (UID: \"4d9f1a15-0cf8-45e2-b4f4-d15d257d7009\") " Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.231918 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3860b8fd-910e-472c-af7e-22021b9c5f42-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.231934 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3860b8fd-910e-472c-af7e-22021b9c5f42-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.234578 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d9f1a15-0cf8-45e2-b4f4-d15d257d7009-scripts" (OuterVolumeSpecName: "scripts") pod "4d9f1a15-0cf8-45e2-b4f4-d15d257d7009" (UID: "4d9f1a15-0cf8-45e2-b4f4-d15d257d7009"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.235650 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d9f1a15-0cf8-45e2-b4f4-d15d257d7009-kube-api-access-d4nvt" (OuterVolumeSpecName: "kube-api-access-d4nvt") pod "4d9f1a15-0cf8-45e2-b4f4-d15d257d7009" (UID: "4d9f1a15-0cf8-45e2-b4f4-d15d257d7009"). InnerVolumeSpecName "kube-api-access-d4nvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.257287 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d9f1a15-0cf8-45e2-b4f4-d15d257d7009-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d9f1a15-0cf8-45e2-b4f4-d15d257d7009" (UID: "4d9f1a15-0cf8-45e2-b4f4-d15d257d7009"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.261959 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d9f1a15-0cf8-45e2-b4f4-d15d257d7009-config-data" (OuterVolumeSpecName: "config-data") pod "4d9f1a15-0cf8-45e2-b4f4-d15d257d7009" (UID: "4d9f1a15-0cf8-45e2-b4f4-d15d257d7009"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.338342 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4nvt\" (UniqueName: \"kubernetes.io/projected/4d9f1a15-0cf8-45e2-b4f4-d15d257d7009-kube-api-access-d4nvt\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.338374 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d9f1a15-0cf8-45e2-b4f4-d15d257d7009-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.338384 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d9f1a15-0cf8-45e2-b4f4-d15d257d7009-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.338391 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d9f1a15-0cf8-45e2-b4f4-d15d257d7009-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.731022 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.732732 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3860b8fd-910e-472c-af7e-22021b9c5f42","Type":"ContainerDied","Data":"e3340d255397d2ecb2720af5f4b9534bfea2b04ddadbeff378958614dc07f4d8"} Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.732805 4813 scope.go:117] "RemoveContainer" containerID="258aee34ba8d549328acc52f2ed0f26867e1376d55703490936bf49c0e56ef3a" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.738398 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-gv4hp" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.738406 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-gv4hp" event={"ID":"4d9f1a15-0cf8-45e2-b4f4-d15d257d7009","Type":"ContainerDied","Data":"1f8ab6c54e878d8569462182cd71bc06222ec6f4d98b90bad9a0702f23ed4ea8"} Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.738448 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f8ab6c54e878d8569462182cd71bc06222ec6f4d98b90bad9a0702f23ed4ea8" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.745035 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"97c18b03-4126-4af3-a790-b7450e54c418","Type":"ContainerDied","Data":"ef16b88288e42da9db6bdcb164fa0b4cefd3dc68b4cd38f775533c557a655fcf"} Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.745117 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.790624 4813 scope.go:117] "RemoveContainer" containerID="67331f1b42d28f68de93ac94725527090c4590e29a20c3f9e232a471759d8b2b" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.798038 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.803885 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.830676 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.841991 4813 scope.go:117] "RemoveContainer" containerID="ea14b6dcb07649c817b1db8a8ab1402ac31c0fe6aa8e7e0e9b4866d917a94aa8" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.844135 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.857175 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:05:32 crc kubenswrapper[4813]: E1206 16:05:32.857559 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97c18b03-4126-4af3-a790-b7450e54c418" containerName="nova-api-api" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.857575 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="97c18b03-4126-4af3-a790-b7450e54c418" containerName="nova-api-api" Dec 06 16:05:32 crc kubenswrapper[4813]: E1206 16:05:32.857587 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3860b8fd-910e-472c-af7e-22021b9c5f42" containerName="proxy-httpd" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.857593 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="3860b8fd-910e-472c-af7e-22021b9c5f42" containerName="proxy-httpd" Dec 06 16:05:32 crc kubenswrapper[4813]: E1206 16:05:32.857604 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3860b8fd-910e-472c-af7e-22021b9c5f42" containerName="ceilometer-notification-agent" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.857611 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="3860b8fd-910e-472c-af7e-22021b9c5f42" containerName="ceilometer-notification-agent" Dec 06 16:05:32 crc kubenswrapper[4813]: E1206 16:05:32.857627 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3860b8fd-910e-472c-af7e-22021b9c5f42" containerName="ceilometer-central-agent" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.857633 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="3860b8fd-910e-472c-af7e-22021b9c5f42" containerName="ceilometer-central-agent" Dec 06 16:05:32 crc kubenswrapper[4813]: E1206 16:05:32.857644 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97c18b03-4126-4af3-a790-b7450e54c418" containerName="nova-api-log" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.857650 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="97c18b03-4126-4af3-a790-b7450e54c418" containerName="nova-api-log" Dec 06 16:05:32 crc kubenswrapper[4813]: E1206 16:05:32.857669 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3860b8fd-910e-472c-af7e-22021b9c5f42" containerName="sg-core" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.857675 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="3860b8fd-910e-472c-af7e-22021b9c5f42" containerName="sg-core" Dec 06 16:05:32 crc kubenswrapper[4813]: E1206 16:05:32.857689 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d9f1a15-0cf8-45e2-b4f4-d15d257d7009" containerName="nova-cell1-conductor-db-sync" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.857695 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d9f1a15-0cf8-45e2-b4f4-d15d257d7009" containerName="nova-cell1-conductor-db-sync" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.857844 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="3860b8fd-910e-472c-af7e-22021b9c5f42" containerName="ceilometer-notification-agent" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.857856 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="3860b8fd-910e-472c-af7e-22021b9c5f42" containerName="sg-core" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.857866 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="3860b8fd-910e-472c-af7e-22021b9c5f42" containerName="proxy-httpd" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.857878 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="3860b8fd-910e-472c-af7e-22021b9c5f42" containerName="ceilometer-central-agent" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.857886 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="97c18b03-4126-4af3-a790-b7450e54c418" containerName="nova-api-log" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.857896 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="97c18b03-4126-4af3-a790-b7450e54c418" containerName="nova-api-api" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.857912 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d9f1a15-0cf8-45e2-b4f4-d15d257d7009" containerName="nova-cell1-conductor-db-sync" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.859445 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.861535 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.861742 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.861742 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.880390 4813 scope.go:117] "RemoveContainer" containerID="2b1742606d409605cb6948ff573d795ff2b4c04a8d7825e8a117f90e873f7472" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.880517 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.891971 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.903367 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.935871 4813 scope.go:117] "RemoveContainer" containerID="56a4caccafc3b2b9f412eb23ce642e5c300d84069300ba7fa0b382fefe96ef39" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.946942 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.950380 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-run-httpd\") pod \"ceilometer-0\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " pod="openstack/ceilometer-0" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.950438 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " pod="openstack/ceilometer-0" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.950792 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/664ed707-8010-46ac-b834-300dc99ca495-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"664ed707-8010-46ac-b834-300dc99ca495\") " pod="openstack/nova-api-0" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.950879 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-log-httpd\") pod \"ceilometer-0\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " pod="openstack/ceilometer-0" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.950949 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-config-data\") pod \"ceilometer-0\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " pod="openstack/ceilometer-0" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.951001 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mfgs\" (UniqueName: \"kubernetes.io/projected/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-kube-api-access-5mfgs\") pod \"ceilometer-0\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " pod="openstack/ceilometer-0" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.951025 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-scripts\") pod \"ceilometer-0\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " pod="openstack/ceilometer-0" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.951105 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " pod="openstack/ceilometer-0" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.951130 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cm8h2\" (UniqueName: \"kubernetes.io/projected/664ed707-8010-46ac-b834-300dc99ca495-kube-api-access-cm8h2\") pod \"nova-api-0\" (UID: \"664ed707-8010-46ac-b834-300dc99ca495\") " pod="openstack/nova-api-0" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.951153 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " pod="openstack/ceilometer-0" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.951173 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/664ed707-8010-46ac-b834-300dc99ca495-logs\") pod \"nova-api-0\" (UID: \"664ed707-8010-46ac-b834-300dc99ca495\") " pod="openstack/nova-api-0" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.951222 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/664ed707-8010-46ac-b834-300dc99ca495-config-data\") pod \"nova-api-0\" (UID: \"664ed707-8010-46ac-b834-300dc99ca495\") " pod="openstack/nova-api-0" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.958147 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.960042 4813 scope.go:117] "RemoveContainer" containerID="1fb6e6ef65034b4e720bedb2ef2b57091bf7d57b3931fb0bac81fbfdf396cb1c" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.960200 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.960654 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.976971 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:05:32 crc kubenswrapper[4813]: I1206 16:05:32.989317 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.053537 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mfgs\" (UniqueName: \"kubernetes.io/projected/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-kube-api-access-5mfgs\") pod \"ceilometer-0\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " pod="openstack/ceilometer-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.053596 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-scripts\") pod \"ceilometer-0\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " pod="openstack/ceilometer-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.053660 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " pod="openstack/ceilometer-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.053683 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cm8h2\" (UniqueName: \"kubernetes.io/projected/664ed707-8010-46ac-b834-300dc99ca495-kube-api-access-cm8h2\") pod \"nova-api-0\" (UID: \"664ed707-8010-46ac-b834-300dc99ca495\") " pod="openstack/nova-api-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.053705 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " pod="openstack/ceilometer-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.053727 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/664ed707-8010-46ac-b834-300dc99ca495-logs\") pod \"nova-api-0\" (UID: \"664ed707-8010-46ac-b834-300dc99ca495\") " pod="openstack/nova-api-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.053764 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/664ed707-8010-46ac-b834-300dc99ca495-config-data\") pod \"nova-api-0\" (UID: \"664ed707-8010-46ac-b834-300dc99ca495\") " pod="openstack/nova-api-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.053808 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-run-httpd\") pod \"ceilometer-0\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " pod="openstack/ceilometer-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.053832 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " pod="openstack/ceilometer-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.053897 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a45c584-458c-4f14-b7f5-50711d94207c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0a45c584-458c-4f14-b7f5-50711d94207c\") " pod="openstack/nova-cell1-conductor-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.053931 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/664ed707-8010-46ac-b834-300dc99ca495-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"664ed707-8010-46ac-b834-300dc99ca495\") " pod="openstack/nova-api-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.053952 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a45c584-458c-4f14-b7f5-50711d94207c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0a45c584-458c-4f14-b7f5-50711d94207c\") " pod="openstack/nova-cell1-conductor-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.053992 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8z6f9\" (UniqueName: \"kubernetes.io/projected/0a45c584-458c-4f14-b7f5-50711d94207c-kube-api-access-8z6f9\") pod \"nova-cell1-conductor-0\" (UID: \"0a45c584-458c-4f14-b7f5-50711d94207c\") " pod="openstack/nova-cell1-conductor-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.054016 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-log-httpd\") pod \"ceilometer-0\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " pod="openstack/ceilometer-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.054059 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-config-data\") pod \"ceilometer-0\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " pod="openstack/ceilometer-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.054245 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/664ed707-8010-46ac-b834-300dc99ca495-logs\") pod \"nova-api-0\" (UID: \"664ed707-8010-46ac-b834-300dc99ca495\") " pod="openstack/nova-api-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.054722 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-run-httpd\") pod \"ceilometer-0\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " pod="openstack/ceilometer-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.054919 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-log-httpd\") pod \"ceilometer-0\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " pod="openstack/ceilometer-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.070459 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/664ed707-8010-46ac-b834-300dc99ca495-config-data\") pod \"nova-api-0\" (UID: \"664ed707-8010-46ac-b834-300dc99ca495\") " pod="openstack/nova-api-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.070506 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " pod="openstack/ceilometer-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.070697 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-config-data\") pod \"ceilometer-0\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " pod="openstack/ceilometer-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.070926 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " pod="openstack/ceilometer-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.071557 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mfgs\" (UniqueName: \"kubernetes.io/projected/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-kube-api-access-5mfgs\") pod \"ceilometer-0\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " pod="openstack/ceilometer-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.071672 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cm8h2\" (UniqueName: \"kubernetes.io/projected/664ed707-8010-46ac-b834-300dc99ca495-kube-api-access-cm8h2\") pod \"nova-api-0\" (UID: \"664ed707-8010-46ac-b834-300dc99ca495\") " pod="openstack/nova-api-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.073298 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/664ed707-8010-46ac-b834-300dc99ca495-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"664ed707-8010-46ac-b834-300dc99ca495\") " pod="openstack/nova-api-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.079035 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-scripts\") pod \"ceilometer-0\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " pod="openstack/ceilometer-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.085527 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " pod="openstack/ceilometer-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.104572 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.104608 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.155633 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a45c584-458c-4f14-b7f5-50711d94207c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0a45c584-458c-4f14-b7f5-50711d94207c\") " pod="openstack/nova-cell1-conductor-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.155909 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a45c584-458c-4f14-b7f5-50711d94207c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0a45c584-458c-4f14-b7f5-50711d94207c\") " pod="openstack/nova-cell1-conductor-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.156028 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8z6f9\" (UniqueName: \"kubernetes.io/projected/0a45c584-458c-4f14-b7f5-50711d94207c-kube-api-access-8z6f9\") pod \"nova-cell1-conductor-0\" (UID: \"0a45c584-458c-4f14-b7f5-50711d94207c\") " pod="openstack/nova-cell1-conductor-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.161333 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a45c584-458c-4f14-b7f5-50711d94207c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0a45c584-458c-4f14-b7f5-50711d94207c\") " pod="openstack/nova-cell1-conductor-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.161533 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a45c584-458c-4f14-b7f5-50711d94207c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0a45c584-458c-4f14-b7f5-50711d94207c\") " pod="openstack/nova-cell1-conductor-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.173907 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8z6f9\" (UniqueName: \"kubernetes.io/projected/0a45c584-458c-4f14-b7f5-50711d94207c-kube-api-access-8z6f9\") pod \"nova-cell1-conductor-0\" (UID: \"0a45c584-458c-4f14-b7f5-50711d94207c\") " pod="openstack/nova-cell1-conductor-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.182183 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.233439 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.273591 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.757810 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.869977 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 16:05:33 crc kubenswrapper[4813]: W1206 16:05:33.879326 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod664ed707_8010_46ac_b834_300dc99ca495.slice/crio-c870d6617c685645cb4d510c7c502ad1aa0461e01a171f4baff9e964400318f5 WatchSource:0}: Error finding container c870d6617c685645cb4d510c7c502ad1aa0461e01a171f4baff9e964400318f5: Status 404 returned error can't find the container with id c870d6617c685645cb4d510c7c502ad1aa0461e01a171f4baff9e964400318f5 Dec 06 16:05:33 crc kubenswrapper[4813]: I1206 16:05:33.889655 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 16:05:33 crc kubenswrapper[4813]: W1206 16:05:33.890908 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a45c584_458c_4f14_b7f5_50711d94207c.slice/crio-6010bd011b701839022a19bdf93521c64a77af9bd8d3bf2f77e4f6555257620a WatchSource:0}: Error finding container 6010bd011b701839022a19bdf93521c64a77af9bd8d3bf2f77e4f6555257620a: Status 404 returned error can't find the container with id 6010bd011b701839022a19bdf93521c64a77af9bd8d3bf2f77e4f6555257620a Dec 06 16:05:34 crc kubenswrapper[4813]: I1206 16:05:34.126962 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 06 16:05:34 crc kubenswrapper[4813]: I1206 16:05:34.496527 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3860b8fd-910e-472c-af7e-22021b9c5f42" path="/var/lib/kubelet/pods/3860b8fd-910e-472c-af7e-22021b9c5f42/volumes" Dec 06 16:05:34 crc kubenswrapper[4813]: I1206 16:05:34.497836 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97c18b03-4126-4af3-a790-b7450e54c418" path="/var/lib/kubelet/pods/97c18b03-4126-4af3-a790-b7450e54c418/volumes" Dec 06 16:05:34 crc kubenswrapper[4813]: I1206 16:05:34.778513 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"664ed707-8010-46ac-b834-300dc99ca495","Type":"ContainerStarted","Data":"f827d4663857d38f36489b24955e94b8ae386ffe84fb35fd352c66ab68f1ee70"} Dec 06 16:05:34 crc kubenswrapper[4813]: I1206 16:05:34.778554 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"664ed707-8010-46ac-b834-300dc99ca495","Type":"ContainerStarted","Data":"55c6a571bb8eeb9eeb5e9240b2cd2157f4490f6de93507251e6a26440d654090"} Dec 06 16:05:34 crc kubenswrapper[4813]: I1206 16:05:34.778565 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"664ed707-8010-46ac-b834-300dc99ca495","Type":"ContainerStarted","Data":"c870d6617c685645cb4d510c7c502ad1aa0461e01a171f4baff9e964400318f5"} Dec 06 16:05:34 crc kubenswrapper[4813]: I1206 16:05:34.781434 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f28c7e49-be3f-45a6-bf8c-2338ffeea04c","Type":"ContainerStarted","Data":"c617fc073ac13c1771155aeae56f8fad360b8f8b93c0471b99e4ba49a98c496d"} Dec 06 16:05:34 crc kubenswrapper[4813]: I1206 16:05:34.781458 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f28c7e49-be3f-45a6-bf8c-2338ffeea04c","Type":"ContainerStarted","Data":"c6efd466287831a04cd801456218a9814d096662abf30b0d8a286f01c456fb9b"} Dec 06 16:05:34 crc kubenswrapper[4813]: I1206 16:05:34.783548 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"0a45c584-458c-4f14-b7f5-50711d94207c","Type":"ContainerStarted","Data":"0288cd0baf0f3f87aadeea347cb485ca6680c7d61e2a646654120f7593dd5ea0"} Dec 06 16:05:34 crc kubenswrapper[4813]: I1206 16:05:34.783571 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"0a45c584-458c-4f14-b7f5-50711d94207c","Type":"ContainerStarted","Data":"6010bd011b701839022a19bdf93521c64a77af9bd8d3bf2f77e4f6555257620a"} Dec 06 16:05:34 crc kubenswrapper[4813]: I1206 16:05:34.783709 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 06 16:05:34 crc kubenswrapper[4813]: I1206 16:05:34.801420 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.80140367 podStartE2EDuration="2.80140367s" podCreationTimestamp="2025-12-06 16:05:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:05:34.794029015 +0000 UTC m=+1174.684908591" watchObservedRunningTime="2025-12-06 16:05:34.80140367 +0000 UTC m=+1174.692283236" Dec 06 16:05:34 crc kubenswrapper[4813]: I1206 16:05:34.817223 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.817206518 podStartE2EDuration="2.817206518s" podCreationTimestamp="2025-12-06 16:05:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:05:34.809792202 +0000 UTC m=+1174.700671778" watchObservedRunningTime="2025-12-06 16:05:34.817206518 +0000 UTC m=+1174.708086094" Dec 06 16:05:35 crc kubenswrapper[4813]: I1206 16:05:35.024449 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 06 16:05:35 crc kubenswrapper[4813]: I1206 16:05:35.794169 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f28c7e49-be3f-45a6-bf8c-2338ffeea04c","Type":"ContainerStarted","Data":"6c80783834070f27988a817dcd2664252bcebe9bdfa6b46b78eb9e00f56ed687"} Dec 06 16:05:36 crc kubenswrapper[4813]: I1206 16:05:36.806594 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f28c7e49-be3f-45a6-bf8c-2338ffeea04c","Type":"ContainerStarted","Data":"840fe6eddf6edc4e0bcfb361a822c95ef6abf2f395802f0c4e8595c40360ae2b"} Dec 06 16:05:37 crc kubenswrapper[4813]: I1206 16:05:37.817695 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f28c7e49-be3f-45a6-bf8c-2338ffeea04c","Type":"ContainerStarted","Data":"67437d23593e0794577a8d91c03747c87d505926e3f315dfb255c38fe86e66ae"} Dec 06 16:05:37 crc kubenswrapper[4813]: I1206 16:05:37.818023 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 16:05:37 crc kubenswrapper[4813]: I1206 16:05:37.844372 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.812626747 podStartE2EDuration="5.844349087s" podCreationTimestamp="2025-12-06 16:05:32 +0000 UTC" firstStartedPulling="2025-12-06 16:05:33.769025631 +0000 UTC m=+1173.659905207" lastFinishedPulling="2025-12-06 16:05:36.800747981 +0000 UTC m=+1176.691627547" observedRunningTime="2025-12-06 16:05:37.839742716 +0000 UTC m=+1177.730622332" watchObservedRunningTime="2025-12-06 16:05:37.844349087 +0000 UTC m=+1177.735228673" Dec 06 16:05:38 crc kubenswrapper[4813]: I1206 16:05:38.106723 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 16:05:38 crc kubenswrapper[4813]: I1206 16:05:38.106787 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 16:05:39 crc kubenswrapper[4813]: I1206 16:05:39.118397 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ff13cc9e-77d0-46a8-9f98-98dd3a75cc99" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 16:05:39 crc kubenswrapper[4813]: I1206 16:05:39.118416 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ff13cc9e-77d0-46a8-9f98-98dd3a75cc99" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 16:05:39 crc kubenswrapper[4813]: I1206 16:05:39.126955 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 06 16:05:39 crc kubenswrapper[4813]: I1206 16:05:39.162718 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 06 16:05:39 crc kubenswrapper[4813]: I1206 16:05:39.898251 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 06 16:05:43 crc kubenswrapper[4813]: I1206 16:05:43.235050 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 16:05:43 crc kubenswrapper[4813]: I1206 16:05:43.236557 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 16:05:43 crc kubenswrapper[4813]: I1206 16:05:43.324966 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 06 16:05:44 crc kubenswrapper[4813]: I1206 16:05:44.317675 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="664ed707-8010-46ac-b834-300dc99ca495" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.197:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 16:05:44 crc kubenswrapper[4813]: I1206 16:05:44.317709 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="664ed707-8010-46ac-b834-300dc99ca495" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.197:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 16:05:48 crc kubenswrapper[4813]: I1206 16:05:48.114008 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 06 16:05:48 crc kubenswrapper[4813]: I1206 16:05:48.117499 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 06 16:05:48 crc kubenswrapper[4813]: I1206 16:05:48.125091 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 06 16:05:48 crc kubenswrapper[4813]: I1206 16:05:48.952226 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 06 16:05:49 crc kubenswrapper[4813]: I1206 16:05:49.428426 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:05:49 crc kubenswrapper[4813]: I1206 16:05:49.428484 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:05:49 crc kubenswrapper[4813]: I1206 16:05:49.817678 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:49 crc kubenswrapper[4813]: I1206 16:05:49.920182 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dlk6p\" (UniqueName: \"kubernetes.io/projected/8cf165d0-93ba-4070-aa79-d659f77c2ffd-kube-api-access-dlk6p\") pod \"8cf165d0-93ba-4070-aa79-d659f77c2ffd\" (UID: \"8cf165d0-93ba-4070-aa79-d659f77c2ffd\") " Dec 06 16:05:49 crc kubenswrapper[4813]: I1206 16:05:49.920328 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cf165d0-93ba-4070-aa79-d659f77c2ffd-config-data\") pod \"8cf165d0-93ba-4070-aa79-d659f77c2ffd\" (UID: \"8cf165d0-93ba-4070-aa79-d659f77c2ffd\") " Dec 06 16:05:49 crc kubenswrapper[4813]: I1206 16:05:49.920472 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cf165d0-93ba-4070-aa79-d659f77c2ffd-combined-ca-bundle\") pod \"8cf165d0-93ba-4070-aa79-d659f77c2ffd\" (UID: \"8cf165d0-93ba-4070-aa79-d659f77c2ffd\") " Dec 06 16:05:49 crc kubenswrapper[4813]: I1206 16:05:49.926704 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cf165d0-93ba-4070-aa79-d659f77c2ffd-kube-api-access-dlk6p" (OuterVolumeSpecName: "kube-api-access-dlk6p") pod "8cf165d0-93ba-4070-aa79-d659f77c2ffd" (UID: "8cf165d0-93ba-4070-aa79-d659f77c2ffd"). InnerVolumeSpecName "kube-api-access-dlk6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:05:49 crc kubenswrapper[4813]: I1206 16:05:49.950642 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cf165d0-93ba-4070-aa79-d659f77c2ffd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8cf165d0-93ba-4070-aa79-d659f77c2ffd" (UID: "8cf165d0-93ba-4070-aa79-d659f77c2ffd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:05:49 crc kubenswrapper[4813]: I1206 16:05:49.952470 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cf165d0-93ba-4070-aa79-d659f77c2ffd-config-data" (OuterVolumeSpecName: "config-data") pod "8cf165d0-93ba-4070-aa79-d659f77c2ffd" (UID: "8cf165d0-93ba-4070-aa79-d659f77c2ffd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:05:49 crc kubenswrapper[4813]: I1206 16:05:49.956928 4813 generic.go:334] "Generic (PLEG): container finished" podID="8cf165d0-93ba-4070-aa79-d659f77c2ffd" containerID="f0f8043b7bb61d5ebdd814f386b66322a981e2053896852e6bbb9d6e8229b396" exitCode=137 Dec 06 16:05:49 crc kubenswrapper[4813]: I1206 16:05:49.956995 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8cf165d0-93ba-4070-aa79-d659f77c2ffd","Type":"ContainerDied","Data":"f0f8043b7bb61d5ebdd814f386b66322a981e2053896852e6bbb9d6e8229b396"} Dec 06 16:05:49 crc kubenswrapper[4813]: I1206 16:05:49.957039 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8cf165d0-93ba-4070-aa79-d659f77c2ffd","Type":"ContainerDied","Data":"d74209a66621bfeda97317d6c56e4a1f05e85252c75d6a6a0856d599371d745b"} Dec 06 16:05:49 crc kubenswrapper[4813]: I1206 16:05:49.957056 4813 scope.go:117] "RemoveContainer" containerID="f0f8043b7bb61d5ebdd814f386b66322a981e2053896852e6bbb9d6e8229b396" Dec 06 16:05:49 crc kubenswrapper[4813]: I1206 16:05:49.957057 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.022613 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cf165d0-93ba-4070-aa79-d659f77c2ffd-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.022638 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cf165d0-93ba-4070-aa79-d659f77c2ffd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.022648 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dlk6p\" (UniqueName: \"kubernetes.io/projected/8cf165d0-93ba-4070-aa79-d659f77c2ffd-kube-api-access-dlk6p\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.059205 4813 scope.go:117] "RemoveContainer" containerID="f0f8043b7bb61d5ebdd814f386b66322a981e2053896852e6bbb9d6e8229b396" Dec 06 16:05:50 crc kubenswrapper[4813]: E1206 16:05:50.060687 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0f8043b7bb61d5ebdd814f386b66322a981e2053896852e6bbb9d6e8229b396\": container with ID starting with f0f8043b7bb61d5ebdd814f386b66322a981e2053896852e6bbb9d6e8229b396 not found: ID does not exist" containerID="f0f8043b7bb61d5ebdd814f386b66322a981e2053896852e6bbb9d6e8229b396" Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.060728 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0f8043b7bb61d5ebdd814f386b66322a981e2053896852e6bbb9d6e8229b396"} err="failed to get container status \"f0f8043b7bb61d5ebdd814f386b66322a981e2053896852e6bbb9d6e8229b396\": rpc error: code = NotFound desc = could not find container \"f0f8043b7bb61d5ebdd814f386b66322a981e2053896852e6bbb9d6e8229b396\": container with ID starting with f0f8043b7bb61d5ebdd814f386b66322a981e2053896852e6bbb9d6e8229b396 not found: ID does not exist" Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.064924 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.072096 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.081178 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 16:05:50 crc kubenswrapper[4813]: E1206 16:05:50.081613 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cf165d0-93ba-4070-aa79-d659f77c2ffd" containerName="nova-cell1-novncproxy-novncproxy" Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.081632 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cf165d0-93ba-4070-aa79-d659f77c2ffd" containerName="nova-cell1-novncproxy-novncproxy" Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.081829 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cf165d0-93ba-4070-aa79-d659f77c2ffd" containerName="nova-cell1-novncproxy-novncproxy" Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.082457 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.084586 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.084809 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.084861 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.105737 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.227491 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9d64e47-211f-4eab-84e9-0caabeff895b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9d64e47-211f-4eab-84e9-0caabeff895b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.227536 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrs5g\" (UniqueName: \"kubernetes.io/projected/e9d64e47-211f-4eab-84e9-0caabeff895b-kube-api-access-zrs5g\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9d64e47-211f-4eab-84e9-0caabeff895b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.227561 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d64e47-211f-4eab-84e9-0caabeff895b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9d64e47-211f-4eab-84e9-0caabeff895b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.227783 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9d64e47-211f-4eab-84e9-0caabeff895b-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9d64e47-211f-4eab-84e9-0caabeff895b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.227905 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9d64e47-211f-4eab-84e9-0caabeff895b-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9d64e47-211f-4eab-84e9-0caabeff895b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.329484 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9d64e47-211f-4eab-84e9-0caabeff895b-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9d64e47-211f-4eab-84e9-0caabeff895b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.329590 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9d64e47-211f-4eab-84e9-0caabeff895b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9d64e47-211f-4eab-84e9-0caabeff895b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.329612 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrs5g\" (UniqueName: \"kubernetes.io/projected/e9d64e47-211f-4eab-84e9-0caabeff895b-kube-api-access-zrs5g\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9d64e47-211f-4eab-84e9-0caabeff895b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.329633 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d64e47-211f-4eab-84e9-0caabeff895b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9d64e47-211f-4eab-84e9-0caabeff895b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.329693 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9d64e47-211f-4eab-84e9-0caabeff895b-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9d64e47-211f-4eab-84e9-0caabeff895b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.333588 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9d64e47-211f-4eab-84e9-0caabeff895b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9d64e47-211f-4eab-84e9-0caabeff895b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.333921 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9d64e47-211f-4eab-84e9-0caabeff895b-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9d64e47-211f-4eab-84e9-0caabeff895b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.334641 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d64e47-211f-4eab-84e9-0caabeff895b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9d64e47-211f-4eab-84e9-0caabeff895b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.334987 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9d64e47-211f-4eab-84e9-0caabeff895b-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9d64e47-211f-4eab-84e9-0caabeff895b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.360974 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrs5g\" (UniqueName: \"kubernetes.io/projected/e9d64e47-211f-4eab-84e9-0caabeff895b-kube-api-access-zrs5g\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9d64e47-211f-4eab-84e9-0caabeff895b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.399725 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.532417 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cf165d0-93ba-4070-aa79-d659f77c2ffd" path="/var/lib/kubelet/pods/8cf165d0-93ba-4070-aa79-d659f77c2ffd/volumes" Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.837812 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 16:05:50 crc kubenswrapper[4813]: W1206 16:05:50.847357 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9d64e47_211f_4eab_84e9_0caabeff895b.slice/crio-888432ef98b3e55d7bee875e8cc1263de2bf6ecdd5f945420bf58e2dc67716ce WatchSource:0}: Error finding container 888432ef98b3e55d7bee875e8cc1263de2bf6ecdd5f945420bf58e2dc67716ce: Status 404 returned error can't find the container with id 888432ef98b3e55d7bee875e8cc1263de2bf6ecdd5f945420bf58e2dc67716ce Dec 06 16:05:50 crc kubenswrapper[4813]: I1206 16:05:50.971511 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e9d64e47-211f-4eab-84e9-0caabeff895b","Type":"ContainerStarted","Data":"888432ef98b3e55d7bee875e8cc1263de2bf6ecdd5f945420bf58e2dc67716ce"} Dec 06 16:05:51 crc kubenswrapper[4813]: I1206 16:05:51.983614 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e9d64e47-211f-4eab-84e9-0caabeff895b","Type":"ContainerStarted","Data":"95587562ff39ee916b0214a5fa914e882e72df27d9a55fcd4295a9c68d60d4d2"} Dec 06 16:05:52 crc kubenswrapper[4813]: I1206 16:05:52.017042 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.017020106 podStartE2EDuration="2.017020106s" podCreationTimestamp="2025-12-06 16:05:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:05:52.010771311 +0000 UTC m=+1191.901650897" watchObservedRunningTime="2025-12-06 16:05:52.017020106 +0000 UTC m=+1191.907899692" Dec 06 16:05:53 crc kubenswrapper[4813]: I1206 16:05:53.240691 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 06 16:05:53 crc kubenswrapper[4813]: I1206 16:05:53.241230 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 06 16:05:53 crc kubenswrapper[4813]: I1206 16:05:53.242661 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 06 16:05:53 crc kubenswrapper[4813]: I1206 16:05:53.242755 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 06 16:05:53 crc kubenswrapper[4813]: I1206 16:05:53.247137 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 06 16:05:53 crc kubenswrapper[4813]: I1206 16:05:53.249541 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 06 16:05:53 crc kubenswrapper[4813]: I1206 16:05:53.510891 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-8rcjh"] Dec 06 16:05:53 crc kubenswrapper[4813]: I1206 16:05:53.514639 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" Dec 06 16:05:53 crc kubenswrapper[4813]: I1206 16:05:53.536721 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-8rcjh"] Dec 06 16:05:53 crc kubenswrapper[4813]: I1206 16:05:53.596679 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-8rcjh\" (UID: \"df5ab4f0-9f97-47a3-89d0-ef1680c215e7\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" Dec 06 16:05:53 crc kubenswrapper[4813]: I1206 16:05:53.596720 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-8rcjh\" (UID: \"df5ab4f0-9f97-47a3-89d0-ef1680c215e7\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" Dec 06 16:05:53 crc kubenswrapper[4813]: I1206 16:05:53.596745 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ws8wh\" (UniqueName: \"kubernetes.io/projected/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-kube-api-access-ws8wh\") pod \"dnsmasq-dns-cd5cbd7b9-8rcjh\" (UID: \"df5ab4f0-9f97-47a3-89d0-ef1680c215e7\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" Dec 06 16:05:53 crc kubenswrapper[4813]: I1206 16:05:53.596773 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-8rcjh\" (UID: \"df5ab4f0-9f97-47a3-89d0-ef1680c215e7\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" Dec 06 16:05:53 crc kubenswrapper[4813]: I1206 16:05:53.596907 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-config\") pod \"dnsmasq-dns-cd5cbd7b9-8rcjh\" (UID: \"df5ab4f0-9f97-47a3-89d0-ef1680c215e7\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" Dec 06 16:05:53 crc kubenswrapper[4813]: I1206 16:05:53.596995 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-8rcjh\" (UID: \"df5ab4f0-9f97-47a3-89d0-ef1680c215e7\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" Dec 06 16:05:53 crc kubenswrapper[4813]: I1206 16:05:53.698934 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-config\") pod \"dnsmasq-dns-cd5cbd7b9-8rcjh\" (UID: \"df5ab4f0-9f97-47a3-89d0-ef1680c215e7\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" Dec 06 16:05:53 crc kubenswrapper[4813]: I1206 16:05:53.699001 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-8rcjh\" (UID: \"df5ab4f0-9f97-47a3-89d0-ef1680c215e7\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" Dec 06 16:05:53 crc kubenswrapper[4813]: I1206 16:05:53.699050 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-8rcjh\" (UID: \"df5ab4f0-9f97-47a3-89d0-ef1680c215e7\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" Dec 06 16:05:53 crc kubenswrapper[4813]: I1206 16:05:53.699070 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-8rcjh\" (UID: \"df5ab4f0-9f97-47a3-89d0-ef1680c215e7\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" Dec 06 16:05:53 crc kubenswrapper[4813]: I1206 16:05:53.699093 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ws8wh\" (UniqueName: \"kubernetes.io/projected/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-kube-api-access-ws8wh\") pod \"dnsmasq-dns-cd5cbd7b9-8rcjh\" (UID: \"df5ab4f0-9f97-47a3-89d0-ef1680c215e7\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" Dec 06 16:05:53 crc kubenswrapper[4813]: I1206 16:05:53.699121 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-8rcjh\" (UID: \"df5ab4f0-9f97-47a3-89d0-ef1680c215e7\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" Dec 06 16:05:53 crc kubenswrapper[4813]: I1206 16:05:53.699874 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-config\") pod \"dnsmasq-dns-cd5cbd7b9-8rcjh\" (UID: \"df5ab4f0-9f97-47a3-89d0-ef1680c215e7\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" Dec 06 16:05:53 crc kubenswrapper[4813]: I1206 16:05:53.699902 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-8rcjh\" (UID: \"df5ab4f0-9f97-47a3-89d0-ef1680c215e7\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" Dec 06 16:05:53 crc kubenswrapper[4813]: I1206 16:05:53.700446 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-8rcjh\" (UID: \"df5ab4f0-9f97-47a3-89d0-ef1680c215e7\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" Dec 06 16:05:53 crc kubenswrapper[4813]: I1206 16:05:53.700519 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-8rcjh\" (UID: \"df5ab4f0-9f97-47a3-89d0-ef1680c215e7\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" Dec 06 16:05:53 crc kubenswrapper[4813]: I1206 16:05:53.701006 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-8rcjh\" (UID: \"df5ab4f0-9f97-47a3-89d0-ef1680c215e7\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" Dec 06 16:05:53 crc kubenswrapper[4813]: I1206 16:05:53.718696 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ws8wh\" (UniqueName: \"kubernetes.io/projected/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-kube-api-access-ws8wh\") pod \"dnsmasq-dns-cd5cbd7b9-8rcjh\" (UID: \"df5ab4f0-9f97-47a3-89d0-ef1680c215e7\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" Dec 06 16:05:53 crc kubenswrapper[4813]: I1206 16:05:53.841437 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" Dec 06 16:05:54 crc kubenswrapper[4813]: I1206 16:05:54.495768 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-8rcjh"] Dec 06 16:05:55 crc kubenswrapper[4813]: I1206 16:05:55.010171 4813 generic.go:334] "Generic (PLEG): container finished" podID="df5ab4f0-9f97-47a3-89d0-ef1680c215e7" containerID="f96efde1a77b21f35c209a2a0aa7f619248974fa7517d0825113a7f5383dc393" exitCode=0 Dec 06 16:05:55 crc kubenswrapper[4813]: I1206 16:05:55.010284 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" event={"ID":"df5ab4f0-9f97-47a3-89d0-ef1680c215e7","Type":"ContainerDied","Data":"f96efde1a77b21f35c209a2a0aa7f619248974fa7517d0825113a7f5383dc393"} Dec 06 16:05:55 crc kubenswrapper[4813]: I1206 16:05:55.010472 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" event={"ID":"df5ab4f0-9f97-47a3-89d0-ef1680c215e7","Type":"ContainerStarted","Data":"25a7109e7cbea43e65fda9df467a29ab5ba06c0c42e7a882c19361e16fc03d23"} Dec 06 16:05:55 crc kubenswrapper[4813]: I1206 16:05:55.400174 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:05:55 crc kubenswrapper[4813]: I1206 16:05:55.617483 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:05:55 crc kubenswrapper[4813]: I1206 16:05:55.617748 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f28c7e49-be3f-45a6-bf8c-2338ffeea04c" containerName="ceilometer-central-agent" containerID="cri-o://c617fc073ac13c1771155aeae56f8fad360b8f8b93c0471b99e4ba49a98c496d" gracePeriod=30 Dec 06 16:05:55 crc kubenswrapper[4813]: I1206 16:05:55.617818 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f28c7e49-be3f-45a6-bf8c-2338ffeea04c" containerName="proxy-httpd" containerID="cri-o://67437d23593e0794577a8d91c03747c87d505926e3f315dfb255c38fe86e66ae" gracePeriod=30 Dec 06 16:05:55 crc kubenswrapper[4813]: I1206 16:05:55.617818 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f28c7e49-be3f-45a6-bf8c-2338ffeea04c" containerName="sg-core" containerID="cri-o://840fe6eddf6edc4e0bcfb361a822c95ef6abf2f395802f0c4e8595c40360ae2b" gracePeriod=30 Dec 06 16:05:55 crc kubenswrapper[4813]: I1206 16:05:55.617835 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f28c7e49-be3f-45a6-bf8c-2338ffeea04c" containerName="ceilometer-notification-agent" containerID="cri-o://6c80783834070f27988a817dcd2664252bcebe9bdfa6b46b78eb9e00f56ed687" gracePeriod=30 Dec 06 16:05:55 crc kubenswrapper[4813]: I1206 16:05:55.813138 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="f28c7e49-be3f-45a6-bf8c-2338ffeea04c" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.021409 4813 generic.go:334] "Generic (PLEG): container finished" podID="f28c7e49-be3f-45a6-bf8c-2338ffeea04c" containerID="840fe6eddf6edc4e0bcfb361a822c95ef6abf2f395802f0c4e8595c40360ae2b" exitCode=2 Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.021467 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f28c7e49-be3f-45a6-bf8c-2338ffeea04c","Type":"ContainerDied","Data":"840fe6eddf6edc4e0bcfb361a822c95ef6abf2f395802f0c4e8595c40360ae2b"} Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.033887 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" event={"ID":"df5ab4f0-9f97-47a3-89d0-ef1680c215e7","Type":"ContainerStarted","Data":"ed00f85126577ceec9342dac944664aacece3dd1fce9976f8bfaf8815e5cd603"} Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.034961 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.056714 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" podStartSLOduration=3.05669773 podStartE2EDuration="3.05669773s" podCreationTimestamp="2025-12-06 16:05:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:05:56.055890128 +0000 UTC m=+1195.946769704" watchObservedRunningTime="2025-12-06 16:05:56.05669773 +0000 UTC m=+1195.947577306" Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.542774 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.542968 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="664ed707-8010-46ac-b834-300dc99ca495" containerName="nova-api-log" containerID="cri-o://55c6a571bb8eeb9eeb5e9240b2cd2157f4490f6de93507251e6a26440d654090" gracePeriod=30 Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.543346 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="664ed707-8010-46ac-b834-300dc99ca495" containerName="nova-api-api" containerID="cri-o://f827d4663857d38f36489b24955e94b8ae386ffe84fb35fd352c66ab68f1ee70" gracePeriod=30 Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.652393 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.756741 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-sg-core-conf-yaml\") pod \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.756807 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mfgs\" (UniqueName: \"kubernetes.io/projected/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-kube-api-access-5mfgs\") pod \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.756953 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-run-httpd\") pod \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.757010 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-log-httpd\") pod \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.757039 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-config-data\") pod \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.757063 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-ceilometer-tls-certs\") pod \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.757079 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-scripts\") pod \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.757110 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-combined-ca-bundle\") pod \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\" (UID: \"f28c7e49-be3f-45a6-bf8c-2338ffeea04c\") " Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.757928 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f28c7e49-be3f-45a6-bf8c-2338ffeea04c" (UID: "f28c7e49-be3f-45a6-bf8c-2338ffeea04c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.758038 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f28c7e49-be3f-45a6-bf8c-2338ffeea04c" (UID: "f28c7e49-be3f-45a6-bf8c-2338ffeea04c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.758048 4813 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.761912 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-kube-api-access-5mfgs" (OuterVolumeSpecName: "kube-api-access-5mfgs") pod "f28c7e49-be3f-45a6-bf8c-2338ffeea04c" (UID: "f28c7e49-be3f-45a6-bf8c-2338ffeea04c"). InnerVolumeSpecName "kube-api-access-5mfgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.764382 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-scripts" (OuterVolumeSpecName: "scripts") pod "f28c7e49-be3f-45a6-bf8c-2338ffeea04c" (UID: "f28c7e49-be3f-45a6-bf8c-2338ffeea04c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.789912 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f28c7e49-be3f-45a6-bf8c-2338ffeea04c" (UID: "f28c7e49-be3f-45a6-bf8c-2338ffeea04c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.818895 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "f28c7e49-be3f-45a6-bf8c-2338ffeea04c" (UID: "f28c7e49-be3f-45a6-bf8c-2338ffeea04c"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.857031 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f28c7e49-be3f-45a6-bf8c-2338ffeea04c" (UID: "f28c7e49-be3f-45a6-bf8c-2338ffeea04c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.860195 4813 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.860222 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mfgs\" (UniqueName: \"kubernetes.io/projected/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-kube-api-access-5mfgs\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.860233 4813 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.860243 4813 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.860251 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.860271 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.864272 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-config-data" (OuterVolumeSpecName: "config-data") pod "f28c7e49-be3f-45a6-bf8c-2338ffeea04c" (UID: "f28c7e49-be3f-45a6-bf8c-2338ffeea04c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:05:56 crc kubenswrapper[4813]: I1206 16:05:56.961751 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f28c7e49-be3f-45a6-bf8c-2338ffeea04c-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.041880 4813 generic.go:334] "Generic (PLEG): container finished" podID="664ed707-8010-46ac-b834-300dc99ca495" containerID="55c6a571bb8eeb9eeb5e9240b2cd2157f4490f6de93507251e6a26440d654090" exitCode=143 Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.041934 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"664ed707-8010-46ac-b834-300dc99ca495","Type":"ContainerDied","Data":"55c6a571bb8eeb9eeb5e9240b2cd2157f4490f6de93507251e6a26440d654090"} Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.044039 4813 generic.go:334] "Generic (PLEG): container finished" podID="f28c7e49-be3f-45a6-bf8c-2338ffeea04c" containerID="67437d23593e0794577a8d91c03747c87d505926e3f315dfb255c38fe86e66ae" exitCode=0 Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.044069 4813 generic.go:334] "Generic (PLEG): container finished" podID="f28c7e49-be3f-45a6-bf8c-2338ffeea04c" containerID="6c80783834070f27988a817dcd2664252bcebe9bdfa6b46b78eb9e00f56ed687" exitCode=0 Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.044078 4813 generic.go:334] "Generic (PLEG): container finished" podID="f28c7e49-be3f-45a6-bf8c-2338ffeea04c" containerID="c617fc073ac13c1771155aeae56f8fad360b8f8b93c0471b99e4ba49a98c496d" exitCode=0 Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.044920 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.051502 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f28c7e49-be3f-45a6-bf8c-2338ffeea04c","Type":"ContainerDied","Data":"67437d23593e0794577a8d91c03747c87d505926e3f315dfb255c38fe86e66ae"} Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.051543 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f28c7e49-be3f-45a6-bf8c-2338ffeea04c","Type":"ContainerDied","Data":"6c80783834070f27988a817dcd2664252bcebe9bdfa6b46b78eb9e00f56ed687"} Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.051554 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f28c7e49-be3f-45a6-bf8c-2338ffeea04c","Type":"ContainerDied","Data":"c617fc073ac13c1771155aeae56f8fad360b8f8b93c0471b99e4ba49a98c496d"} Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.051563 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f28c7e49-be3f-45a6-bf8c-2338ffeea04c","Type":"ContainerDied","Data":"c6efd466287831a04cd801456218a9814d096662abf30b0d8a286f01c456fb9b"} Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.051580 4813 scope.go:117] "RemoveContainer" containerID="67437d23593e0794577a8d91c03747c87d505926e3f315dfb255c38fe86e66ae" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.078621 4813 scope.go:117] "RemoveContainer" containerID="840fe6eddf6edc4e0bcfb361a822c95ef6abf2f395802f0c4e8595c40360ae2b" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.079826 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.092058 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.099786 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:05:57 crc kubenswrapper[4813]: E1206 16:05:57.100131 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f28c7e49-be3f-45a6-bf8c-2338ffeea04c" containerName="sg-core" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.100143 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f28c7e49-be3f-45a6-bf8c-2338ffeea04c" containerName="sg-core" Dec 06 16:05:57 crc kubenswrapper[4813]: E1206 16:05:57.100160 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f28c7e49-be3f-45a6-bf8c-2338ffeea04c" containerName="proxy-httpd" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.100166 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f28c7e49-be3f-45a6-bf8c-2338ffeea04c" containerName="proxy-httpd" Dec 06 16:05:57 crc kubenswrapper[4813]: E1206 16:05:57.100187 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f28c7e49-be3f-45a6-bf8c-2338ffeea04c" containerName="ceilometer-notification-agent" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.100193 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f28c7e49-be3f-45a6-bf8c-2338ffeea04c" containerName="ceilometer-notification-agent" Dec 06 16:05:57 crc kubenswrapper[4813]: E1206 16:05:57.100214 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f28c7e49-be3f-45a6-bf8c-2338ffeea04c" containerName="ceilometer-central-agent" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.100220 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f28c7e49-be3f-45a6-bf8c-2338ffeea04c" containerName="ceilometer-central-agent" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.100387 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f28c7e49-be3f-45a6-bf8c-2338ffeea04c" containerName="ceilometer-notification-agent" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.100395 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f28c7e49-be3f-45a6-bf8c-2338ffeea04c" containerName="ceilometer-central-agent" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.100415 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f28c7e49-be3f-45a6-bf8c-2338ffeea04c" containerName="sg-core" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.100422 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f28c7e49-be3f-45a6-bf8c-2338ffeea04c" containerName="proxy-httpd" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.102328 4813 scope.go:117] "RemoveContainer" containerID="6c80783834070f27988a817dcd2664252bcebe9bdfa6b46b78eb9e00f56ed687" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.103397 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.105051 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.106109 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.106392 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.132015 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.138782 4813 scope.go:117] "RemoveContainer" containerID="c617fc073ac13c1771155aeae56f8fad360b8f8b93c0471b99e4ba49a98c496d" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.157087 4813 scope.go:117] "RemoveContainer" containerID="67437d23593e0794577a8d91c03747c87d505926e3f315dfb255c38fe86e66ae" Dec 06 16:05:57 crc kubenswrapper[4813]: E1206 16:05:57.158209 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67437d23593e0794577a8d91c03747c87d505926e3f315dfb255c38fe86e66ae\": container with ID starting with 67437d23593e0794577a8d91c03747c87d505926e3f315dfb255c38fe86e66ae not found: ID does not exist" containerID="67437d23593e0794577a8d91c03747c87d505926e3f315dfb255c38fe86e66ae" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.158389 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67437d23593e0794577a8d91c03747c87d505926e3f315dfb255c38fe86e66ae"} err="failed to get container status \"67437d23593e0794577a8d91c03747c87d505926e3f315dfb255c38fe86e66ae\": rpc error: code = NotFound desc = could not find container \"67437d23593e0794577a8d91c03747c87d505926e3f315dfb255c38fe86e66ae\": container with ID starting with 67437d23593e0794577a8d91c03747c87d505926e3f315dfb255c38fe86e66ae not found: ID does not exist" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.158470 4813 scope.go:117] "RemoveContainer" containerID="840fe6eddf6edc4e0bcfb361a822c95ef6abf2f395802f0c4e8595c40360ae2b" Dec 06 16:05:57 crc kubenswrapper[4813]: E1206 16:05:57.158759 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"840fe6eddf6edc4e0bcfb361a822c95ef6abf2f395802f0c4e8595c40360ae2b\": container with ID starting with 840fe6eddf6edc4e0bcfb361a822c95ef6abf2f395802f0c4e8595c40360ae2b not found: ID does not exist" containerID="840fe6eddf6edc4e0bcfb361a822c95ef6abf2f395802f0c4e8595c40360ae2b" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.158831 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"840fe6eddf6edc4e0bcfb361a822c95ef6abf2f395802f0c4e8595c40360ae2b"} err="failed to get container status \"840fe6eddf6edc4e0bcfb361a822c95ef6abf2f395802f0c4e8595c40360ae2b\": rpc error: code = NotFound desc = could not find container \"840fe6eddf6edc4e0bcfb361a822c95ef6abf2f395802f0c4e8595c40360ae2b\": container with ID starting with 840fe6eddf6edc4e0bcfb361a822c95ef6abf2f395802f0c4e8595c40360ae2b not found: ID does not exist" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.158902 4813 scope.go:117] "RemoveContainer" containerID="6c80783834070f27988a817dcd2664252bcebe9bdfa6b46b78eb9e00f56ed687" Dec 06 16:05:57 crc kubenswrapper[4813]: E1206 16:05:57.159182 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c80783834070f27988a817dcd2664252bcebe9bdfa6b46b78eb9e00f56ed687\": container with ID starting with 6c80783834070f27988a817dcd2664252bcebe9bdfa6b46b78eb9e00f56ed687 not found: ID does not exist" containerID="6c80783834070f27988a817dcd2664252bcebe9bdfa6b46b78eb9e00f56ed687" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.159287 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c80783834070f27988a817dcd2664252bcebe9bdfa6b46b78eb9e00f56ed687"} err="failed to get container status \"6c80783834070f27988a817dcd2664252bcebe9bdfa6b46b78eb9e00f56ed687\": rpc error: code = NotFound desc = could not find container \"6c80783834070f27988a817dcd2664252bcebe9bdfa6b46b78eb9e00f56ed687\": container with ID starting with 6c80783834070f27988a817dcd2664252bcebe9bdfa6b46b78eb9e00f56ed687 not found: ID does not exist" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.159364 4813 scope.go:117] "RemoveContainer" containerID="c617fc073ac13c1771155aeae56f8fad360b8f8b93c0471b99e4ba49a98c496d" Dec 06 16:05:57 crc kubenswrapper[4813]: E1206 16:05:57.159771 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c617fc073ac13c1771155aeae56f8fad360b8f8b93c0471b99e4ba49a98c496d\": container with ID starting with c617fc073ac13c1771155aeae56f8fad360b8f8b93c0471b99e4ba49a98c496d not found: ID does not exist" containerID="c617fc073ac13c1771155aeae56f8fad360b8f8b93c0471b99e4ba49a98c496d" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.159860 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c617fc073ac13c1771155aeae56f8fad360b8f8b93c0471b99e4ba49a98c496d"} err="failed to get container status \"c617fc073ac13c1771155aeae56f8fad360b8f8b93c0471b99e4ba49a98c496d\": rpc error: code = NotFound desc = could not find container \"c617fc073ac13c1771155aeae56f8fad360b8f8b93c0471b99e4ba49a98c496d\": container with ID starting with c617fc073ac13c1771155aeae56f8fad360b8f8b93c0471b99e4ba49a98c496d not found: ID does not exist" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.159937 4813 scope.go:117] "RemoveContainer" containerID="67437d23593e0794577a8d91c03747c87d505926e3f315dfb255c38fe86e66ae" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.160222 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67437d23593e0794577a8d91c03747c87d505926e3f315dfb255c38fe86e66ae"} err="failed to get container status \"67437d23593e0794577a8d91c03747c87d505926e3f315dfb255c38fe86e66ae\": rpc error: code = NotFound desc = could not find container \"67437d23593e0794577a8d91c03747c87d505926e3f315dfb255c38fe86e66ae\": container with ID starting with 67437d23593e0794577a8d91c03747c87d505926e3f315dfb255c38fe86e66ae not found: ID does not exist" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.160327 4813 scope.go:117] "RemoveContainer" containerID="840fe6eddf6edc4e0bcfb361a822c95ef6abf2f395802f0c4e8595c40360ae2b" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.160570 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"840fe6eddf6edc4e0bcfb361a822c95ef6abf2f395802f0c4e8595c40360ae2b"} err="failed to get container status \"840fe6eddf6edc4e0bcfb361a822c95ef6abf2f395802f0c4e8595c40360ae2b\": rpc error: code = NotFound desc = could not find container \"840fe6eddf6edc4e0bcfb361a822c95ef6abf2f395802f0c4e8595c40360ae2b\": container with ID starting with 840fe6eddf6edc4e0bcfb361a822c95ef6abf2f395802f0c4e8595c40360ae2b not found: ID does not exist" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.160669 4813 scope.go:117] "RemoveContainer" containerID="6c80783834070f27988a817dcd2664252bcebe9bdfa6b46b78eb9e00f56ed687" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.161147 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c80783834070f27988a817dcd2664252bcebe9bdfa6b46b78eb9e00f56ed687"} err="failed to get container status \"6c80783834070f27988a817dcd2664252bcebe9bdfa6b46b78eb9e00f56ed687\": rpc error: code = NotFound desc = could not find container \"6c80783834070f27988a817dcd2664252bcebe9bdfa6b46b78eb9e00f56ed687\": container with ID starting with 6c80783834070f27988a817dcd2664252bcebe9bdfa6b46b78eb9e00f56ed687 not found: ID does not exist" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.161290 4813 scope.go:117] "RemoveContainer" containerID="c617fc073ac13c1771155aeae56f8fad360b8f8b93c0471b99e4ba49a98c496d" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.161556 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c617fc073ac13c1771155aeae56f8fad360b8f8b93c0471b99e4ba49a98c496d"} err="failed to get container status \"c617fc073ac13c1771155aeae56f8fad360b8f8b93c0471b99e4ba49a98c496d\": rpc error: code = NotFound desc = could not find container \"c617fc073ac13c1771155aeae56f8fad360b8f8b93c0471b99e4ba49a98c496d\": container with ID starting with c617fc073ac13c1771155aeae56f8fad360b8f8b93c0471b99e4ba49a98c496d not found: ID does not exist" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.161657 4813 scope.go:117] "RemoveContainer" containerID="67437d23593e0794577a8d91c03747c87d505926e3f315dfb255c38fe86e66ae" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.162034 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67437d23593e0794577a8d91c03747c87d505926e3f315dfb255c38fe86e66ae"} err="failed to get container status \"67437d23593e0794577a8d91c03747c87d505926e3f315dfb255c38fe86e66ae\": rpc error: code = NotFound desc = could not find container \"67437d23593e0794577a8d91c03747c87d505926e3f315dfb255c38fe86e66ae\": container with ID starting with 67437d23593e0794577a8d91c03747c87d505926e3f315dfb255c38fe86e66ae not found: ID does not exist" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.162128 4813 scope.go:117] "RemoveContainer" containerID="840fe6eddf6edc4e0bcfb361a822c95ef6abf2f395802f0c4e8595c40360ae2b" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.162460 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"840fe6eddf6edc4e0bcfb361a822c95ef6abf2f395802f0c4e8595c40360ae2b"} err="failed to get container status \"840fe6eddf6edc4e0bcfb361a822c95ef6abf2f395802f0c4e8595c40360ae2b\": rpc error: code = NotFound desc = could not find container \"840fe6eddf6edc4e0bcfb361a822c95ef6abf2f395802f0c4e8595c40360ae2b\": container with ID starting with 840fe6eddf6edc4e0bcfb361a822c95ef6abf2f395802f0c4e8595c40360ae2b not found: ID does not exist" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.162531 4813 scope.go:117] "RemoveContainer" containerID="6c80783834070f27988a817dcd2664252bcebe9bdfa6b46b78eb9e00f56ed687" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.163419 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c80783834070f27988a817dcd2664252bcebe9bdfa6b46b78eb9e00f56ed687"} err="failed to get container status \"6c80783834070f27988a817dcd2664252bcebe9bdfa6b46b78eb9e00f56ed687\": rpc error: code = NotFound desc = could not find container \"6c80783834070f27988a817dcd2664252bcebe9bdfa6b46b78eb9e00f56ed687\": container with ID starting with 6c80783834070f27988a817dcd2664252bcebe9bdfa6b46b78eb9e00f56ed687 not found: ID does not exist" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.163442 4813 scope.go:117] "RemoveContainer" containerID="c617fc073ac13c1771155aeae56f8fad360b8f8b93c0471b99e4ba49a98c496d" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.163620 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c617fc073ac13c1771155aeae56f8fad360b8f8b93c0471b99e4ba49a98c496d"} err="failed to get container status \"c617fc073ac13c1771155aeae56f8fad360b8f8b93c0471b99e4ba49a98c496d\": rpc error: code = NotFound desc = could not find container \"c617fc073ac13c1771155aeae56f8fad360b8f8b93c0471b99e4ba49a98c496d\": container with ID starting with c617fc073ac13c1771155aeae56f8fad360b8f8b93c0471b99e4ba49a98c496d not found: ID does not exist" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.168544 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-run-httpd\") pod \"ceilometer-0\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " pod="openstack/ceilometer-0" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.168686 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " pod="openstack/ceilometer-0" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.168709 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " pod="openstack/ceilometer-0" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.168750 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-scripts\") pod \"ceilometer-0\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " pod="openstack/ceilometer-0" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.168780 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " pod="openstack/ceilometer-0" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.168804 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkd5w\" (UniqueName: \"kubernetes.io/projected/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-kube-api-access-tkd5w\") pod \"ceilometer-0\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " pod="openstack/ceilometer-0" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.168839 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-config-data\") pod \"ceilometer-0\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " pod="openstack/ceilometer-0" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.168869 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-log-httpd\") pod \"ceilometer-0\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " pod="openstack/ceilometer-0" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.269813 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-config-data\") pod \"ceilometer-0\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " pod="openstack/ceilometer-0" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.270076 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-log-httpd\") pod \"ceilometer-0\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " pod="openstack/ceilometer-0" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.270194 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-run-httpd\") pod \"ceilometer-0\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " pod="openstack/ceilometer-0" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.270333 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " pod="openstack/ceilometer-0" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.270410 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " pod="openstack/ceilometer-0" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.270489 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-scripts\") pod \"ceilometer-0\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " pod="openstack/ceilometer-0" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.270570 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " pod="openstack/ceilometer-0" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.270639 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-log-httpd\") pod \"ceilometer-0\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " pod="openstack/ceilometer-0" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.270649 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkd5w\" (UniqueName: \"kubernetes.io/projected/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-kube-api-access-tkd5w\") pod \"ceilometer-0\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " pod="openstack/ceilometer-0" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.271185 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-run-httpd\") pod \"ceilometer-0\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " pod="openstack/ceilometer-0" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.274499 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " pod="openstack/ceilometer-0" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.274604 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-scripts\") pod \"ceilometer-0\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " pod="openstack/ceilometer-0" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.275308 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-config-data\") pod \"ceilometer-0\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " pod="openstack/ceilometer-0" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.276370 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " pod="openstack/ceilometer-0" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.278995 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " pod="openstack/ceilometer-0" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.297905 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkd5w\" (UniqueName: \"kubernetes.io/projected/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-kube-api-access-tkd5w\") pod \"ceilometer-0\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " pod="openstack/ceilometer-0" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.423597 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.905784 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:05:57 crc kubenswrapper[4813]: I1206 16:05:57.925337 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:05:58 crc kubenswrapper[4813]: I1206 16:05:58.064164 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2427ee4-0c95-4db6-8464-5dfb6ca5b735","Type":"ContainerStarted","Data":"eabd5e8e7603b5410432a85bf4cae2c7e86d54a09c7c7665e15f4eb7b0d13ad0"} Dec 06 16:05:58 crc kubenswrapper[4813]: I1206 16:05:58.496605 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f28c7e49-be3f-45a6-bf8c-2338ffeea04c" path="/var/lib/kubelet/pods/f28c7e49-be3f-45a6-bf8c-2338ffeea04c/volumes" Dec 06 16:05:59 crc kubenswrapper[4813]: I1206 16:05:59.074584 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2427ee4-0c95-4db6-8464-5dfb6ca5b735","Type":"ContainerStarted","Data":"9463eecd2eae2d259f47b1a2d83299570b169d062240eef1ee73b2b16b45e34d"} Dec 06 16:06:00 crc kubenswrapper[4813]: I1206 16:06:00.030009 4813 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 16:06:00 crc kubenswrapper[4813]: I1206 16:06:00.092115 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2427ee4-0c95-4db6-8464-5dfb6ca5b735","Type":"ContainerStarted","Data":"58d5c1e461520886787eb0bf24d1eff79728d7b2f18e13b818d8232c60bd12b0"} Dec 06 16:06:00 crc kubenswrapper[4813]: I1206 16:06:00.092157 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2427ee4-0c95-4db6-8464-5dfb6ca5b735","Type":"ContainerStarted","Data":"44fa0cc80e79448d04e08d3a9f4a860a0c9d85ed90f50271b8c10117e06a813d"} Dec 06 16:06:00 crc kubenswrapper[4813]: I1206 16:06:00.095013 4813 generic.go:334] "Generic (PLEG): container finished" podID="664ed707-8010-46ac-b834-300dc99ca495" containerID="f827d4663857d38f36489b24955e94b8ae386ffe84fb35fd352c66ab68f1ee70" exitCode=0 Dec 06 16:06:00 crc kubenswrapper[4813]: I1206 16:06:00.095042 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"664ed707-8010-46ac-b834-300dc99ca495","Type":"ContainerDied","Data":"f827d4663857d38f36489b24955e94b8ae386ffe84fb35fd352c66ab68f1ee70"} Dec 06 16:06:00 crc kubenswrapper[4813]: I1206 16:06:00.118657 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 16:06:00 crc kubenswrapper[4813]: I1206 16:06:00.223842 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cm8h2\" (UniqueName: \"kubernetes.io/projected/664ed707-8010-46ac-b834-300dc99ca495-kube-api-access-cm8h2\") pod \"664ed707-8010-46ac-b834-300dc99ca495\" (UID: \"664ed707-8010-46ac-b834-300dc99ca495\") " Dec 06 16:06:00 crc kubenswrapper[4813]: I1206 16:06:00.223899 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/664ed707-8010-46ac-b834-300dc99ca495-logs\") pod \"664ed707-8010-46ac-b834-300dc99ca495\" (UID: \"664ed707-8010-46ac-b834-300dc99ca495\") " Dec 06 16:06:00 crc kubenswrapper[4813]: I1206 16:06:00.224006 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/664ed707-8010-46ac-b834-300dc99ca495-config-data\") pod \"664ed707-8010-46ac-b834-300dc99ca495\" (UID: \"664ed707-8010-46ac-b834-300dc99ca495\") " Dec 06 16:06:00 crc kubenswrapper[4813]: I1206 16:06:00.224093 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/664ed707-8010-46ac-b834-300dc99ca495-combined-ca-bundle\") pod \"664ed707-8010-46ac-b834-300dc99ca495\" (UID: \"664ed707-8010-46ac-b834-300dc99ca495\") " Dec 06 16:06:00 crc kubenswrapper[4813]: I1206 16:06:00.225125 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/664ed707-8010-46ac-b834-300dc99ca495-logs" (OuterVolumeSpecName: "logs") pod "664ed707-8010-46ac-b834-300dc99ca495" (UID: "664ed707-8010-46ac-b834-300dc99ca495"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:06:00 crc kubenswrapper[4813]: I1206 16:06:00.231082 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/664ed707-8010-46ac-b834-300dc99ca495-kube-api-access-cm8h2" (OuterVolumeSpecName: "kube-api-access-cm8h2") pod "664ed707-8010-46ac-b834-300dc99ca495" (UID: "664ed707-8010-46ac-b834-300dc99ca495"). InnerVolumeSpecName "kube-api-access-cm8h2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:06:00 crc kubenswrapper[4813]: I1206 16:06:00.311317 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/664ed707-8010-46ac-b834-300dc99ca495-config-data" (OuterVolumeSpecName: "config-data") pod "664ed707-8010-46ac-b834-300dc99ca495" (UID: "664ed707-8010-46ac-b834-300dc99ca495"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:06:00 crc kubenswrapper[4813]: I1206 16:06:00.313429 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/664ed707-8010-46ac-b834-300dc99ca495-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "664ed707-8010-46ac-b834-300dc99ca495" (UID: "664ed707-8010-46ac-b834-300dc99ca495"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:06:00 crc kubenswrapper[4813]: I1206 16:06:00.327356 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/664ed707-8010-46ac-b834-300dc99ca495-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:00 crc kubenswrapper[4813]: I1206 16:06:00.327382 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/664ed707-8010-46ac-b834-300dc99ca495-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:00 crc kubenswrapper[4813]: I1206 16:06:00.327394 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cm8h2\" (UniqueName: \"kubernetes.io/projected/664ed707-8010-46ac-b834-300dc99ca495-kube-api-access-cm8h2\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:00 crc kubenswrapper[4813]: I1206 16:06:00.327402 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/664ed707-8010-46ac-b834-300dc99ca495-logs\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:00 crc kubenswrapper[4813]: I1206 16:06:00.402309 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:06:00 crc kubenswrapper[4813]: I1206 16:06:00.444600 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.104556 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"664ed707-8010-46ac-b834-300dc99ca495","Type":"ContainerDied","Data":"c870d6617c685645cb4d510c7c502ad1aa0461e01a171f4baff9e964400318f5"} Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.104855 4813 scope.go:117] "RemoveContainer" containerID="f827d4663857d38f36489b24955e94b8ae386ffe84fb35fd352c66ab68f1ee70" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.104603 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.136577 4813 scope.go:117] "RemoveContainer" containerID="55c6a571bb8eeb9eeb5e9240b2cd2157f4490f6de93507251e6a26440d654090" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.136711 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.156409 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.172624 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 06 16:06:01 crc kubenswrapper[4813]: E1206 16:06:01.172974 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="664ed707-8010-46ac-b834-300dc99ca495" containerName="nova-api-api" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.172990 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="664ed707-8010-46ac-b834-300dc99ca495" containerName="nova-api-api" Dec 06 16:06:01 crc kubenswrapper[4813]: E1206 16:06:01.173010 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="664ed707-8010-46ac-b834-300dc99ca495" containerName="nova-api-log" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.173017 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="664ed707-8010-46ac-b834-300dc99ca495" containerName="nova-api-log" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.173179 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="664ed707-8010-46ac-b834-300dc99ca495" containerName="nova-api-api" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.173203 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="664ed707-8010-46ac-b834-300dc99ca495" containerName="nova-api-log" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.174086 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.179764 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.180033 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.180143 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.197078 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.218858 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.245292 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce376ede-55df-470d-bb45-584bad422181-public-tls-certs\") pod \"nova-api-0\" (UID: \"ce376ede-55df-470d-bb45-584bad422181\") " pod="openstack/nova-api-0" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.245342 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfqzd\" (UniqueName: \"kubernetes.io/projected/ce376ede-55df-470d-bb45-584bad422181-kube-api-access-xfqzd\") pod \"nova-api-0\" (UID: \"ce376ede-55df-470d-bb45-584bad422181\") " pod="openstack/nova-api-0" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.245393 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce376ede-55df-470d-bb45-584bad422181-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ce376ede-55df-470d-bb45-584bad422181\") " pod="openstack/nova-api-0" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.245432 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce376ede-55df-470d-bb45-584bad422181-config-data\") pod \"nova-api-0\" (UID: \"ce376ede-55df-470d-bb45-584bad422181\") " pod="openstack/nova-api-0" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.245449 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce376ede-55df-470d-bb45-584bad422181-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ce376ede-55df-470d-bb45-584bad422181\") " pod="openstack/nova-api-0" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.245485 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce376ede-55df-470d-bb45-584bad422181-logs\") pod \"nova-api-0\" (UID: \"ce376ede-55df-470d-bb45-584bad422181\") " pod="openstack/nova-api-0" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.348269 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce376ede-55df-470d-bb45-584bad422181-logs\") pod \"nova-api-0\" (UID: \"ce376ede-55df-470d-bb45-584bad422181\") " pod="openstack/nova-api-0" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.348532 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce376ede-55df-470d-bb45-584bad422181-public-tls-certs\") pod \"nova-api-0\" (UID: \"ce376ede-55df-470d-bb45-584bad422181\") " pod="openstack/nova-api-0" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.348564 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfqzd\" (UniqueName: \"kubernetes.io/projected/ce376ede-55df-470d-bb45-584bad422181-kube-api-access-xfqzd\") pod \"nova-api-0\" (UID: \"ce376ede-55df-470d-bb45-584bad422181\") " pod="openstack/nova-api-0" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.348608 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce376ede-55df-470d-bb45-584bad422181-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ce376ede-55df-470d-bb45-584bad422181\") " pod="openstack/nova-api-0" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.348679 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce376ede-55df-470d-bb45-584bad422181-config-data\") pod \"nova-api-0\" (UID: \"ce376ede-55df-470d-bb45-584bad422181\") " pod="openstack/nova-api-0" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.348696 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce376ede-55df-470d-bb45-584bad422181-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ce376ede-55df-470d-bb45-584bad422181\") " pod="openstack/nova-api-0" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.351892 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce376ede-55df-470d-bb45-584bad422181-logs\") pod \"nova-api-0\" (UID: \"ce376ede-55df-470d-bb45-584bad422181\") " pod="openstack/nova-api-0" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.354838 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce376ede-55df-470d-bb45-584bad422181-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ce376ede-55df-470d-bb45-584bad422181\") " pod="openstack/nova-api-0" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.354899 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce376ede-55df-470d-bb45-584bad422181-config-data\") pod \"nova-api-0\" (UID: \"ce376ede-55df-470d-bb45-584bad422181\") " pod="openstack/nova-api-0" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.360890 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce376ede-55df-470d-bb45-584bad422181-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ce376ede-55df-470d-bb45-584bad422181\") " pod="openstack/nova-api-0" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.364608 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce376ede-55df-470d-bb45-584bad422181-public-tls-certs\") pod \"nova-api-0\" (UID: \"ce376ede-55df-470d-bb45-584bad422181\") " pod="openstack/nova-api-0" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.372709 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfqzd\" (UniqueName: \"kubernetes.io/projected/ce376ede-55df-470d-bb45-584bad422181-kube-api-access-xfqzd\") pod \"nova-api-0\" (UID: \"ce376ede-55df-470d-bb45-584bad422181\") " pod="openstack/nova-api-0" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.484904 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-nq29n"] Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.486090 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-nq29n" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.488450 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.488778 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.503085 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-nq29n"] Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.535369 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.552145 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b77b30e-c026-408e-a869-0bdc2c1d650a-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-nq29n\" (UID: \"1b77b30e-c026-408e-a869-0bdc2c1d650a\") " pod="openstack/nova-cell1-cell-mapping-nq29n" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.552312 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b77b30e-c026-408e-a869-0bdc2c1d650a-scripts\") pod \"nova-cell1-cell-mapping-nq29n\" (UID: \"1b77b30e-c026-408e-a869-0bdc2c1d650a\") " pod="openstack/nova-cell1-cell-mapping-nq29n" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.552417 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpcbk\" (UniqueName: \"kubernetes.io/projected/1b77b30e-c026-408e-a869-0bdc2c1d650a-kube-api-access-qpcbk\") pod \"nova-cell1-cell-mapping-nq29n\" (UID: \"1b77b30e-c026-408e-a869-0bdc2c1d650a\") " pod="openstack/nova-cell1-cell-mapping-nq29n" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.552497 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b77b30e-c026-408e-a869-0bdc2c1d650a-config-data\") pod \"nova-cell1-cell-mapping-nq29n\" (UID: \"1b77b30e-c026-408e-a869-0bdc2c1d650a\") " pod="openstack/nova-cell1-cell-mapping-nq29n" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.656232 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b77b30e-c026-408e-a869-0bdc2c1d650a-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-nq29n\" (UID: \"1b77b30e-c026-408e-a869-0bdc2c1d650a\") " pod="openstack/nova-cell1-cell-mapping-nq29n" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.657092 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b77b30e-c026-408e-a869-0bdc2c1d650a-scripts\") pod \"nova-cell1-cell-mapping-nq29n\" (UID: \"1b77b30e-c026-408e-a869-0bdc2c1d650a\") " pod="openstack/nova-cell1-cell-mapping-nq29n" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.657152 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpcbk\" (UniqueName: \"kubernetes.io/projected/1b77b30e-c026-408e-a869-0bdc2c1d650a-kube-api-access-qpcbk\") pod \"nova-cell1-cell-mapping-nq29n\" (UID: \"1b77b30e-c026-408e-a869-0bdc2c1d650a\") " pod="openstack/nova-cell1-cell-mapping-nq29n" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.657177 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b77b30e-c026-408e-a869-0bdc2c1d650a-config-data\") pod \"nova-cell1-cell-mapping-nq29n\" (UID: \"1b77b30e-c026-408e-a869-0bdc2c1d650a\") " pod="openstack/nova-cell1-cell-mapping-nq29n" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.661068 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b77b30e-c026-408e-a869-0bdc2c1d650a-scripts\") pod \"nova-cell1-cell-mapping-nq29n\" (UID: \"1b77b30e-c026-408e-a869-0bdc2c1d650a\") " pod="openstack/nova-cell1-cell-mapping-nq29n" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.666662 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b77b30e-c026-408e-a869-0bdc2c1d650a-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-nq29n\" (UID: \"1b77b30e-c026-408e-a869-0bdc2c1d650a\") " pod="openstack/nova-cell1-cell-mapping-nq29n" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.677851 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b77b30e-c026-408e-a869-0bdc2c1d650a-config-data\") pod \"nova-cell1-cell-mapping-nq29n\" (UID: \"1b77b30e-c026-408e-a869-0bdc2c1d650a\") " pod="openstack/nova-cell1-cell-mapping-nq29n" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.682652 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpcbk\" (UniqueName: \"kubernetes.io/projected/1b77b30e-c026-408e-a869-0bdc2c1d650a-kube-api-access-qpcbk\") pod \"nova-cell1-cell-mapping-nq29n\" (UID: \"1b77b30e-c026-408e-a869-0bdc2c1d650a\") " pod="openstack/nova-cell1-cell-mapping-nq29n" Dec 06 16:06:01 crc kubenswrapper[4813]: I1206 16:06:01.824346 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-nq29n" Dec 06 16:06:02 crc kubenswrapper[4813]: I1206 16:06:02.053546 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 16:06:02 crc kubenswrapper[4813]: W1206 16:06:02.057918 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce376ede_55df_470d_bb45_584bad422181.slice/crio-eac9506246601ea3f2cfe0294677c9d8a4f5c475271530791b7d676cf61ac4e0 WatchSource:0}: Error finding container eac9506246601ea3f2cfe0294677c9d8a4f5c475271530791b7d676cf61ac4e0: Status 404 returned error can't find the container with id eac9506246601ea3f2cfe0294677c9d8a4f5c475271530791b7d676cf61ac4e0 Dec 06 16:06:02 crc kubenswrapper[4813]: I1206 16:06:02.115299 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce376ede-55df-470d-bb45-584bad422181","Type":"ContainerStarted","Data":"eac9506246601ea3f2cfe0294677c9d8a4f5c475271530791b7d676cf61ac4e0"} Dec 06 16:06:02 crc kubenswrapper[4813]: I1206 16:06:02.129811 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2427ee4-0c95-4db6-8464-5dfb6ca5b735","Type":"ContainerStarted","Data":"075ffa95057b853bcb4f65a5a5c2f7b30e9e142a70e3367526f85e261a81e3fb"} Dec 06 16:06:02 crc kubenswrapper[4813]: I1206 16:06:02.129888 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d2427ee4-0c95-4db6-8464-5dfb6ca5b735" containerName="ceilometer-central-agent" containerID="cri-o://9463eecd2eae2d259f47b1a2d83299570b169d062240eef1ee73b2b16b45e34d" gracePeriod=30 Dec 06 16:06:02 crc kubenswrapper[4813]: I1206 16:06:02.129980 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d2427ee4-0c95-4db6-8464-5dfb6ca5b735" containerName="ceilometer-notification-agent" containerID="cri-o://44fa0cc80e79448d04e08d3a9f4a860a0c9d85ed90f50271b8c10117e06a813d" gracePeriod=30 Dec 06 16:06:02 crc kubenswrapper[4813]: I1206 16:06:02.129984 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d2427ee4-0c95-4db6-8464-5dfb6ca5b735" containerName="proxy-httpd" containerID="cri-o://075ffa95057b853bcb4f65a5a5c2f7b30e9e142a70e3367526f85e261a81e3fb" gracePeriod=30 Dec 06 16:06:02 crc kubenswrapper[4813]: I1206 16:06:02.129984 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d2427ee4-0c95-4db6-8464-5dfb6ca5b735" containerName="sg-core" containerID="cri-o://58d5c1e461520886787eb0bf24d1eff79728d7b2f18e13b818d8232c60bd12b0" gracePeriod=30 Dec 06 16:06:02 crc kubenswrapper[4813]: I1206 16:06:02.130111 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 16:06:02 crc kubenswrapper[4813]: I1206 16:06:02.154984 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.855663206 podStartE2EDuration="5.154968562s" podCreationTimestamp="2025-12-06 16:05:57 +0000 UTC" firstStartedPulling="2025-12-06 16:05:57.932105173 +0000 UTC m=+1197.822984749" lastFinishedPulling="2025-12-06 16:06:01.231410529 +0000 UTC m=+1201.122290105" observedRunningTime="2025-12-06 16:06:02.152569258 +0000 UTC m=+1202.043448824" watchObservedRunningTime="2025-12-06 16:06:02.154968562 +0000 UTC m=+1202.045848138" Dec 06 16:06:02 crc kubenswrapper[4813]: I1206 16:06:02.262987 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-nq29n"] Dec 06 16:06:02 crc kubenswrapper[4813]: I1206 16:06:02.514444 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="664ed707-8010-46ac-b834-300dc99ca495" path="/var/lib/kubelet/pods/664ed707-8010-46ac-b834-300dc99ca495/volumes" Dec 06 16:06:03 crc kubenswrapper[4813]: I1206 16:06:03.153482 4813 generic.go:334] "Generic (PLEG): container finished" podID="d2427ee4-0c95-4db6-8464-5dfb6ca5b735" containerID="075ffa95057b853bcb4f65a5a5c2f7b30e9e142a70e3367526f85e261a81e3fb" exitCode=0 Dec 06 16:06:03 crc kubenswrapper[4813]: I1206 16:06:03.153542 4813 generic.go:334] "Generic (PLEG): container finished" podID="d2427ee4-0c95-4db6-8464-5dfb6ca5b735" containerID="58d5c1e461520886787eb0bf24d1eff79728d7b2f18e13b818d8232c60bd12b0" exitCode=2 Dec 06 16:06:03 crc kubenswrapper[4813]: I1206 16:06:03.153532 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2427ee4-0c95-4db6-8464-5dfb6ca5b735","Type":"ContainerDied","Data":"075ffa95057b853bcb4f65a5a5c2f7b30e9e142a70e3367526f85e261a81e3fb"} Dec 06 16:06:03 crc kubenswrapper[4813]: I1206 16:06:03.153606 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2427ee4-0c95-4db6-8464-5dfb6ca5b735","Type":"ContainerDied","Data":"58d5c1e461520886787eb0bf24d1eff79728d7b2f18e13b818d8232c60bd12b0"} Dec 06 16:06:03 crc kubenswrapper[4813]: I1206 16:06:03.153630 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2427ee4-0c95-4db6-8464-5dfb6ca5b735","Type":"ContainerDied","Data":"44fa0cc80e79448d04e08d3a9f4a860a0c9d85ed90f50271b8c10117e06a813d"} Dec 06 16:06:03 crc kubenswrapper[4813]: I1206 16:06:03.153555 4813 generic.go:334] "Generic (PLEG): container finished" podID="d2427ee4-0c95-4db6-8464-5dfb6ca5b735" containerID="44fa0cc80e79448d04e08d3a9f4a860a0c9d85ed90f50271b8c10117e06a813d" exitCode=0 Dec 06 16:06:03 crc kubenswrapper[4813]: I1206 16:06:03.157618 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-nq29n" event={"ID":"1b77b30e-c026-408e-a869-0bdc2c1d650a","Type":"ContainerStarted","Data":"987af28cdd40354bf772b307c353116f8cff0959e35004737b938db61dea3de7"} Dec 06 16:06:03 crc kubenswrapper[4813]: I1206 16:06:03.157700 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-nq29n" event={"ID":"1b77b30e-c026-408e-a869-0bdc2c1d650a","Type":"ContainerStarted","Data":"4e0a8390537505e4577dc6fcbc7a9de5d385eeba6362d1c5c5a6a816c80d3cb2"} Dec 06 16:06:03 crc kubenswrapper[4813]: I1206 16:06:03.165192 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce376ede-55df-470d-bb45-584bad422181","Type":"ContainerStarted","Data":"556d98347b33cbb2eb093b02b6f4195f88bbdebb4af27a70012ab51a941682e0"} Dec 06 16:06:03 crc kubenswrapper[4813]: I1206 16:06:03.165254 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce376ede-55df-470d-bb45-584bad422181","Type":"ContainerStarted","Data":"6353afba07a44d441a2043bba5066bfa6bd9fa1ce62aa5d74d84e9969fddbf13"} Dec 06 16:06:03 crc kubenswrapper[4813]: I1206 16:06:03.173123 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-nq29n" podStartSLOduration=2.173099694 podStartE2EDuration="2.173099694s" podCreationTimestamp="2025-12-06 16:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:06:03.172590881 +0000 UTC m=+1203.063470457" watchObservedRunningTime="2025-12-06 16:06:03.173099694 +0000 UTC m=+1203.063979300" Dec 06 16:06:03 crc kubenswrapper[4813]: I1206 16:06:03.202145 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.202123272 podStartE2EDuration="2.202123272s" podCreationTimestamp="2025-12-06 16:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:06:03.191404408 +0000 UTC m=+1203.082284004" watchObservedRunningTime="2025-12-06 16:06:03.202123272 +0000 UTC m=+1203.093002858" Dec 06 16:06:03 crc kubenswrapper[4813]: I1206 16:06:03.843601 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" Dec 06 16:06:03 crc kubenswrapper[4813]: I1206 16:06:03.934588 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-4w48r"] Dec 06 16:06:03 crc kubenswrapper[4813]: I1206 16:06:03.935094 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bccf8f775-4w48r" podUID="826c1ef8-3395-4fde-9ef3-fac3d0486f41" containerName="dnsmasq-dns" containerID="cri-o://75d9e1196f0a65ef60f3a05343344987a95b4755a57be6d5479cf51b6d56ea04" gracePeriod=10 Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.180105 4813 generic.go:334] "Generic (PLEG): container finished" podID="826c1ef8-3395-4fde-9ef3-fac3d0486f41" containerID="75d9e1196f0a65ef60f3a05343344987a95b4755a57be6d5479cf51b6d56ea04" exitCode=0 Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.181209 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-4w48r" event={"ID":"826c1ef8-3395-4fde-9ef3-fac3d0486f41","Type":"ContainerDied","Data":"75d9e1196f0a65ef60f3a05343344987a95b4755a57be6d5479cf51b6d56ea04"} Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.488803 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-4w48r" Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.643033 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-ovsdbserver-sb\") pod \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\" (UID: \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\") " Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.643105 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-dns-svc\") pod \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\" (UID: \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\") " Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.643122 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-dns-swift-storage-0\") pod \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\" (UID: \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\") " Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.643271 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-ovsdbserver-nb\") pod \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\" (UID: \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\") " Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.643308 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-config\") pod \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\" (UID: \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\") " Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.643345 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvshl\" (UniqueName: \"kubernetes.io/projected/826c1ef8-3395-4fde-9ef3-fac3d0486f41-kube-api-access-lvshl\") pod \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\" (UID: \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\") " Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.675564 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/826c1ef8-3395-4fde-9ef3-fac3d0486f41-kube-api-access-lvshl" (OuterVolumeSpecName: "kube-api-access-lvshl") pod "826c1ef8-3395-4fde-9ef3-fac3d0486f41" (UID: "826c1ef8-3395-4fde-9ef3-fac3d0486f41"). InnerVolumeSpecName "kube-api-access-lvshl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.706671 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.720045 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "826c1ef8-3395-4fde-9ef3-fac3d0486f41" (UID: "826c1ef8-3395-4fde-9ef3-fac3d0486f41"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.751848 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "826c1ef8-3395-4fde-9ef3-fac3d0486f41" (UID: "826c1ef8-3395-4fde-9ef3-fac3d0486f41"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.751982 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-config" (OuterVolumeSpecName: "config") pod "826c1ef8-3395-4fde-9ef3-fac3d0486f41" (UID: "826c1ef8-3395-4fde-9ef3-fac3d0486f41"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.752187 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "826c1ef8-3395-4fde-9ef3-fac3d0486f41" (UID: "826c1ef8-3395-4fde-9ef3-fac3d0486f41"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.752755 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-ovsdbserver-sb\") pod \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\" (UID: \"826c1ef8-3395-4fde-9ef3-fac3d0486f41\") " Dec 06 16:06:04 crc kubenswrapper[4813]: W1206 16:06:04.752903 4813 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/826c1ef8-3395-4fde-9ef3-fac3d0486f41/volumes/kubernetes.io~configmap/ovsdbserver-sb Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.752921 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "826c1ef8-3395-4fde-9ef3-fac3d0486f41" (UID: "826c1ef8-3395-4fde-9ef3-fac3d0486f41"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.753485 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.753552 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-config\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.753617 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvshl\" (UniqueName: \"kubernetes.io/projected/826c1ef8-3395-4fde-9ef3-fac3d0486f41-kube-api-access-lvshl\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.753680 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.753927 4813 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.772773 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "826c1ef8-3395-4fde-9ef3-fac3d0486f41" (UID: "826c1ef8-3395-4fde-9ef3-fac3d0486f41"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.855158 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-config-data\") pod \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.855459 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-ceilometer-tls-certs\") pod \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.855498 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-combined-ca-bundle\") pod \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.855535 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkd5w\" (UniqueName: \"kubernetes.io/projected/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-kube-api-access-tkd5w\") pod \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.855568 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-sg-core-conf-yaml\") pod \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.855643 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-log-httpd\") pod \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.855684 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-run-httpd\") pod \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.855772 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-scripts\") pod \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\" (UID: \"d2427ee4-0c95-4db6-8464-5dfb6ca5b735\") " Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.856147 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/826c1ef8-3395-4fde-9ef3-fac3d0486f41-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.857031 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d2427ee4-0c95-4db6-8464-5dfb6ca5b735" (UID: "d2427ee4-0c95-4db6-8464-5dfb6ca5b735"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.857134 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d2427ee4-0c95-4db6-8464-5dfb6ca5b735" (UID: "d2427ee4-0c95-4db6-8464-5dfb6ca5b735"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.859487 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-scripts" (OuterVolumeSpecName: "scripts") pod "d2427ee4-0c95-4db6-8464-5dfb6ca5b735" (UID: "d2427ee4-0c95-4db6-8464-5dfb6ca5b735"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.860337 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-kube-api-access-tkd5w" (OuterVolumeSpecName: "kube-api-access-tkd5w") pod "d2427ee4-0c95-4db6-8464-5dfb6ca5b735" (UID: "d2427ee4-0c95-4db6-8464-5dfb6ca5b735"). InnerVolumeSpecName "kube-api-access-tkd5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.883047 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d2427ee4-0c95-4db6-8464-5dfb6ca5b735" (UID: "d2427ee4-0c95-4db6-8464-5dfb6ca5b735"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.906850 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "d2427ee4-0c95-4db6-8464-5dfb6ca5b735" (UID: "d2427ee4-0c95-4db6-8464-5dfb6ca5b735"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.957888 4813 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.957914 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkd5w\" (UniqueName: \"kubernetes.io/projected/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-kube-api-access-tkd5w\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.957923 4813 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.957931 4813 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.957940 4813 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.957950 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.961743 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-config-data" (OuterVolumeSpecName: "config-data") pod "d2427ee4-0c95-4db6-8464-5dfb6ca5b735" (UID: "d2427ee4-0c95-4db6-8464-5dfb6ca5b735"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:06:04 crc kubenswrapper[4813]: I1206 16:06:04.970456 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d2427ee4-0c95-4db6-8464-5dfb6ca5b735" (UID: "d2427ee4-0c95-4db6-8464-5dfb6ca5b735"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.059591 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.059624 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2427ee4-0c95-4db6-8464-5dfb6ca5b735-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.190048 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-4w48r" event={"ID":"826c1ef8-3395-4fde-9ef3-fac3d0486f41","Type":"ContainerDied","Data":"d90f7d1b3433ad16104b32ede7b6a9de735b7a1a6394f3f8717e678e9e42d176"} Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.190101 4813 scope.go:117] "RemoveContainer" containerID="75d9e1196f0a65ef60f3a05343344987a95b4755a57be6d5479cf51b6d56ea04" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.190233 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-4w48r" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.193820 4813 generic.go:334] "Generic (PLEG): container finished" podID="d2427ee4-0c95-4db6-8464-5dfb6ca5b735" containerID="9463eecd2eae2d259f47b1a2d83299570b169d062240eef1ee73b2b16b45e34d" exitCode=0 Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.193861 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2427ee4-0c95-4db6-8464-5dfb6ca5b735","Type":"ContainerDied","Data":"9463eecd2eae2d259f47b1a2d83299570b169d062240eef1ee73b2b16b45e34d"} Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.193887 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2427ee4-0c95-4db6-8464-5dfb6ca5b735","Type":"ContainerDied","Data":"eabd5e8e7603b5410432a85bf4cae2c7e86d54a09c7c7665e15f4eb7b0d13ad0"} Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.193948 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.227012 4813 scope.go:117] "RemoveContainer" containerID="f8f601193777045356d0473da46e26c63b720ca68e50e80769ff6583b64a8c07" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.236353 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-4w48r"] Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.254422 4813 scope.go:117] "RemoveContainer" containerID="075ffa95057b853bcb4f65a5a5c2f7b30e9e142a70e3367526f85e261a81e3fb" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.255408 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-4w48r"] Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.269323 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.286337 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.315247 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:06:05 crc kubenswrapper[4813]: E1206 16:06:05.315705 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="826c1ef8-3395-4fde-9ef3-fac3d0486f41" containerName="dnsmasq-dns" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.315729 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="826c1ef8-3395-4fde-9ef3-fac3d0486f41" containerName="dnsmasq-dns" Dec 06 16:06:05 crc kubenswrapper[4813]: E1206 16:06:05.315749 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="826c1ef8-3395-4fde-9ef3-fac3d0486f41" containerName="init" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.315756 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="826c1ef8-3395-4fde-9ef3-fac3d0486f41" containerName="init" Dec 06 16:06:05 crc kubenswrapper[4813]: E1206 16:06:05.315768 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2427ee4-0c95-4db6-8464-5dfb6ca5b735" containerName="ceilometer-central-agent" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.315775 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2427ee4-0c95-4db6-8464-5dfb6ca5b735" containerName="ceilometer-central-agent" Dec 06 16:06:05 crc kubenswrapper[4813]: E1206 16:06:05.315789 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2427ee4-0c95-4db6-8464-5dfb6ca5b735" containerName="ceilometer-notification-agent" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.315796 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2427ee4-0c95-4db6-8464-5dfb6ca5b735" containerName="ceilometer-notification-agent" Dec 06 16:06:05 crc kubenswrapper[4813]: E1206 16:06:05.315804 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2427ee4-0c95-4db6-8464-5dfb6ca5b735" containerName="sg-core" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.315811 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2427ee4-0c95-4db6-8464-5dfb6ca5b735" containerName="sg-core" Dec 06 16:06:05 crc kubenswrapper[4813]: E1206 16:06:05.315824 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2427ee4-0c95-4db6-8464-5dfb6ca5b735" containerName="proxy-httpd" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.315833 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2427ee4-0c95-4db6-8464-5dfb6ca5b735" containerName="proxy-httpd" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.315993 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="826c1ef8-3395-4fde-9ef3-fac3d0486f41" containerName="dnsmasq-dns" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.316014 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2427ee4-0c95-4db6-8464-5dfb6ca5b735" containerName="ceilometer-central-agent" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.316021 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2427ee4-0c95-4db6-8464-5dfb6ca5b735" containerName="sg-core" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.316032 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2427ee4-0c95-4db6-8464-5dfb6ca5b735" containerName="ceilometer-notification-agent" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.316039 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2427ee4-0c95-4db6-8464-5dfb6ca5b735" containerName="proxy-httpd" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.317674 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.321642 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.329511 4813 scope.go:117] "RemoveContainer" containerID="58d5c1e461520886787eb0bf24d1eff79728d7b2f18e13b818d8232c60bd12b0" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.329584 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.329669 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.329705 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.352957 4813 scope.go:117] "RemoveContainer" containerID="44fa0cc80e79448d04e08d3a9f4a860a0c9d85ed90f50271b8c10117e06a813d" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.386457 4813 scope.go:117] "RemoveContainer" containerID="9463eecd2eae2d259f47b1a2d83299570b169d062240eef1ee73b2b16b45e34d" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.411523 4813 scope.go:117] "RemoveContainer" containerID="075ffa95057b853bcb4f65a5a5c2f7b30e9e142a70e3367526f85e261a81e3fb" Dec 06 16:06:05 crc kubenswrapper[4813]: E1206 16:06:05.411941 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"075ffa95057b853bcb4f65a5a5c2f7b30e9e142a70e3367526f85e261a81e3fb\": container with ID starting with 075ffa95057b853bcb4f65a5a5c2f7b30e9e142a70e3367526f85e261a81e3fb not found: ID does not exist" containerID="075ffa95057b853bcb4f65a5a5c2f7b30e9e142a70e3367526f85e261a81e3fb" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.411984 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"075ffa95057b853bcb4f65a5a5c2f7b30e9e142a70e3367526f85e261a81e3fb"} err="failed to get container status \"075ffa95057b853bcb4f65a5a5c2f7b30e9e142a70e3367526f85e261a81e3fb\": rpc error: code = NotFound desc = could not find container \"075ffa95057b853bcb4f65a5a5c2f7b30e9e142a70e3367526f85e261a81e3fb\": container with ID starting with 075ffa95057b853bcb4f65a5a5c2f7b30e9e142a70e3367526f85e261a81e3fb not found: ID does not exist" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.412009 4813 scope.go:117] "RemoveContainer" containerID="58d5c1e461520886787eb0bf24d1eff79728d7b2f18e13b818d8232c60bd12b0" Dec 06 16:06:05 crc kubenswrapper[4813]: E1206 16:06:05.412274 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58d5c1e461520886787eb0bf24d1eff79728d7b2f18e13b818d8232c60bd12b0\": container with ID starting with 58d5c1e461520886787eb0bf24d1eff79728d7b2f18e13b818d8232c60bd12b0 not found: ID does not exist" containerID="58d5c1e461520886787eb0bf24d1eff79728d7b2f18e13b818d8232c60bd12b0" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.412302 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58d5c1e461520886787eb0bf24d1eff79728d7b2f18e13b818d8232c60bd12b0"} err="failed to get container status \"58d5c1e461520886787eb0bf24d1eff79728d7b2f18e13b818d8232c60bd12b0\": rpc error: code = NotFound desc = could not find container \"58d5c1e461520886787eb0bf24d1eff79728d7b2f18e13b818d8232c60bd12b0\": container with ID starting with 58d5c1e461520886787eb0bf24d1eff79728d7b2f18e13b818d8232c60bd12b0 not found: ID does not exist" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.412327 4813 scope.go:117] "RemoveContainer" containerID="44fa0cc80e79448d04e08d3a9f4a860a0c9d85ed90f50271b8c10117e06a813d" Dec 06 16:06:05 crc kubenswrapper[4813]: E1206 16:06:05.412712 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44fa0cc80e79448d04e08d3a9f4a860a0c9d85ed90f50271b8c10117e06a813d\": container with ID starting with 44fa0cc80e79448d04e08d3a9f4a860a0c9d85ed90f50271b8c10117e06a813d not found: ID does not exist" containerID="44fa0cc80e79448d04e08d3a9f4a860a0c9d85ed90f50271b8c10117e06a813d" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.412737 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44fa0cc80e79448d04e08d3a9f4a860a0c9d85ed90f50271b8c10117e06a813d"} err="failed to get container status \"44fa0cc80e79448d04e08d3a9f4a860a0c9d85ed90f50271b8c10117e06a813d\": rpc error: code = NotFound desc = could not find container \"44fa0cc80e79448d04e08d3a9f4a860a0c9d85ed90f50271b8c10117e06a813d\": container with ID starting with 44fa0cc80e79448d04e08d3a9f4a860a0c9d85ed90f50271b8c10117e06a813d not found: ID does not exist" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.412751 4813 scope.go:117] "RemoveContainer" containerID="9463eecd2eae2d259f47b1a2d83299570b169d062240eef1ee73b2b16b45e34d" Dec 06 16:06:05 crc kubenswrapper[4813]: E1206 16:06:05.413268 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9463eecd2eae2d259f47b1a2d83299570b169d062240eef1ee73b2b16b45e34d\": container with ID starting with 9463eecd2eae2d259f47b1a2d83299570b169d062240eef1ee73b2b16b45e34d not found: ID does not exist" containerID="9463eecd2eae2d259f47b1a2d83299570b169d062240eef1ee73b2b16b45e34d" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.413326 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9463eecd2eae2d259f47b1a2d83299570b169d062240eef1ee73b2b16b45e34d"} err="failed to get container status \"9463eecd2eae2d259f47b1a2d83299570b169d062240eef1ee73b2b16b45e34d\": rpc error: code = NotFound desc = could not find container \"9463eecd2eae2d259f47b1a2d83299570b169d062240eef1ee73b2b16b45e34d\": container with ID starting with 9463eecd2eae2d259f47b1a2d83299570b169d062240eef1ee73b2b16b45e34d not found: ID does not exist" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.473039 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pph7c\" (UniqueName: \"kubernetes.io/projected/66613ee6-b0b7-48fc-adb6-86bab4940550-kube-api-access-pph7c\") pod \"ceilometer-0\" (UID: \"66613ee6-b0b7-48fc-adb6-86bab4940550\") " pod="openstack/ceilometer-0" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.473325 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66613ee6-b0b7-48fc-adb6-86bab4940550-config-data\") pod \"ceilometer-0\" (UID: \"66613ee6-b0b7-48fc-adb6-86bab4940550\") " pod="openstack/ceilometer-0" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.473354 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66613ee6-b0b7-48fc-adb6-86bab4940550-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"66613ee6-b0b7-48fc-adb6-86bab4940550\") " pod="openstack/ceilometer-0" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.473405 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66613ee6-b0b7-48fc-adb6-86bab4940550-run-httpd\") pod \"ceilometer-0\" (UID: \"66613ee6-b0b7-48fc-adb6-86bab4940550\") " pod="openstack/ceilometer-0" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.473428 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/66613ee6-b0b7-48fc-adb6-86bab4940550-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"66613ee6-b0b7-48fc-adb6-86bab4940550\") " pod="openstack/ceilometer-0" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.473444 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66613ee6-b0b7-48fc-adb6-86bab4940550-log-httpd\") pod \"ceilometer-0\" (UID: \"66613ee6-b0b7-48fc-adb6-86bab4940550\") " pod="openstack/ceilometer-0" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.473517 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66613ee6-b0b7-48fc-adb6-86bab4940550-scripts\") pod \"ceilometer-0\" (UID: \"66613ee6-b0b7-48fc-adb6-86bab4940550\") " pod="openstack/ceilometer-0" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.473546 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/66613ee6-b0b7-48fc-adb6-86bab4940550-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"66613ee6-b0b7-48fc-adb6-86bab4940550\") " pod="openstack/ceilometer-0" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.575222 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/66613ee6-b0b7-48fc-adb6-86bab4940550-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"66613ee6-b0b7-48fc-adb6-86bab4940550\") " pod="openstack/ceilometer-0" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.575344 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pph7c\" (UniqueName: \"kubernetes.io/projected/66613ee6-b0b7-48fc-adb6-86bab4940550-kube-api-access-pph7c\") pod \"ceilometer-0\" (UID: \"66613ee6-b0b7-48fc-adb6-86bab4940550\") " pod="openstack/ceilometer-0" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.575406 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66613ee6-b0b7-48fc-adb6-86bab4940550-config-data\") pod \"ceilometer-0\" (UID: \"66613ee6-b0b7-48fc-adb6-86bab4940550\") " pod="openstack/ceilometer-0" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.575440 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66613ee6-b0b7-48fc-adb6-86bab4940550-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"66613ee6-b0b7-48fc-adb6-86bab4940550\") " pod="openstack/ceilometer-0" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.575502 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66613ee6-b0b7-48fc-adb6-86bab4940550-run-httpd\") pod \"ceilometer-0\" (UID: \"66613ee6-b0b7-48fc-adb6-86bab4940550\") " pod="openstack/ceilometer-0" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.575533 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/66613ee6-b0b7-48fc-adb6-86bab4940550-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"66613ee6-b0b7-48fc-adb6-86bab4940550\") " pod="openstack/ceilometer-0" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.575548 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66613ee6-b0b7-48fc-adb6-86bab4940550-log-httpd\") pod \"ceilometer-0\" (UID: \"66613ee6-b0b7-48fc-adb6-86bab4940550\") " pod="openstack/ceilometer-0" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.575594 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66613ee6-b0b7-48fc-adb6-86bab4940550-scripts\") pod \"ceilometer-0\" (UID: \"66613ee6-b0b7-48fc-adb6-86bab4940550\") " pod="openstack/ceilometer-0" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.576302 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66613ee6-b0b7-48fc-adb6-86bab4940550-run-httpd\") pod \"ceilometer-0\" (UID: \"66613ee6-b0b7-48fc-adb6-86bab4940550\") " pod="openstack/ceilometer-0" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.576893 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66613ee6-b0b7-48fc-adb6-86bab4940550-log-httpd\") pod \"ceilometer-0\" (UID: \"66613ee6-b0b7-48fc-adb6-86bab4940550\") " pod="openstack/ceilometer-0" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.590411 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66613ee6-b0b7-48fc-adb6-86bab4940550-scripts\") pod \"ceilometer-0\" (UID: \"66613ee6-b0b7-48fc-adb6-86bab4940550\") " pod="openstack/ceilometer-0" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.595720 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/66613ee6-b0b7-48fc-adb6-86bab4940550-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"66613ee6-b0b7-48fc-adb6-86bab4940550\") " pod="openstack/ceilometer-0" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.597045 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/66613ee6-b0b7-48fc-adb6-86bab4940550-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"66613ee6-b0b7-48fc-adb6-86bab4940550\") " pod="openstack/ceilometer-0" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.606787 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66613ee6-b0b7-48fc-adb6-86bab4940550-config-data\") pod \"ceilometer-0\" (UID: \"66613ee6-b0b7-48fc-adb6-86bab4940550\") " pod="openstack/ceilometer-0" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.613519 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pph7c\" (UniqueName: \"kubernetes.io/projected/66613ee6-b0b7-48fc-adb6-86bab4940550-kube-api-access-pph7c\") pod \"ceilometer-0\" (UID: \"66613ee6-b0b7-48fc-adb6-86bab4940550\") " pod="openstack/ceilometer-0" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.616241 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66613ee6-b0b7-48fc-adb6-86bab4940550-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"66613ee6-b0b7-48fc-adb6-86bab4940550\") " pod="openstack/ceilometer-0" Dec 06 16:06:05 crc kubenswrapper[4813]: I1206 16:06:05.639074 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 16:06:06 crc kubenswrapper[4813]: I1206 16:06:06.106405 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 16:06:06 crc kubenswrapper[4813]: I1206 16:06:06.204849 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66613ee6-b0b7-48fc-adb6-86bab4940550","Type":"ContainerStarted","Data":"dd347f69e4d189a4975b5262d44fe81e882932c276bad3b05e9ec9de355fa23f"} Dec 06 16:06:06 crc kubenswrapper[4813]: I1206 16:06:06.497740 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="826c1ef8-3395-4fde-9ef3-fac3d0486f41" path="/var/lib/kubelet/pods/826c1ef8-3395-4fde-9ef3-fac3d0486f41/volumes" Dec 06 16:06:06 crc kubenswrapper[4813]: I1206 16:06:06.498645 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2427ee4-0c95-4db6-8464-5dfb6ca5b735" path="/var/lib/kubelet/pods/d2427ee4-0c95-4db6-8464-5dfb6ca5b735/volumes" Dec 06 16:06:07 crc kubenswrapper[4813]: I1206 16:06:07.217592 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66613ee6-b0b7-48fc-adb6-86bab4940550","Type":"ContainerStarted","Data":"5c0c2a14208542244a6803a8718f5f9a6eee4d61a6e8f2a6129c45901c671f5a"} Dec 06 16:06:08 crc kubenswrapper[4813]: I1206 16:06:08.226086 4813 generic.go:334] "Generic (PLEG): container finished" podID="1b77b30e-c026-408e-a869-0bdc2c1d650a" containerID="987af28cdd40354bf772b307c353116f8cff0959e35004737b938db61dea3de7" exitCode=0 Dec 06 16:06:08 crc kubenswrapper[4813]: I1206 16:06:08.226411 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-nq29n" event={"ID":"1b77b30e-c026-408e-a869-0bdc2c1d650a","Type":"ContainerDied","Data":"987af28cdd40354bf772b307c353116f8cff0959e35004737b938db61dea3de7"} Dec 06 16:06:08 crc kubenswrapper[4813]: I1206 16:06:08.228998 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66613ee6-b0b7-48fc-adb6-86bab4940550","Type":"ContainerStarted","Data":"99234a67bdc1bf9df92b877a7d34101b902242b552944741562608efac90a872"} Dec 06 16:06:08 crc kubenswrapper[4813]: I1206 16:06:08.229022 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66613ee6-b0b7-48fc-adb6-86bab4940550","Type":"ContainerStarted","Data":"f132aae468323b3fb8c644d68803f2d2dd000359527ad806d98dd7cba5a8fb12"} Dec 06 16:06:09 crc kubenswrapper[4813]: I1206 16:06:09.610576 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-nq29n" Dec 06 16:06:09 crc kubenswrapper[4813]: I1206 16:06:09.756102 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b77b30e-c026-408e-a869-0bdc2c1d650a-combined-ca-bundle\") pod \"1b77b30e-c026-408e-a869-0bdc2c1d650a\" (UID: \"1b77b30e-c026-408e-a869-0bdc2c1d650a\") " Dec 06 16:06:09 crc kubenswrapper[4813]: I1206 16:06:09.756241 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b77b30e-c026-408e-a869-0bdc2c1d650a-config-data\") pod \"1b77b30e-c026-408e-a869-0bdc2c1d650a\" (UID: \"1b77b30e-c026-408e-a869-0bdc2c1d650a\") " Dec 06 16:06:09 crc kubenswrapper[4813]: I1206 16:06:09.756278 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpcbk\" (UniqueName: \"kubernetes.io/projected/1b77b30e-c026-408e-a869-0bdc2c1d650a-kube-api-access-qpcbk\") pod \"1b77b30e-c026-408e-a869-0bdc2c1d650a\" (UID: \"1b77b30e-c026-408e-a869-0bdc2c1d650a\") " Dec 06 16:06:09 crc kubenswrapper[4813]: I1206 16:06:09.756325 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b77b30e-c026-408e-a869-0bdc2c1d650a-scripts\") pod \"1b77b30e-c026-408e-a869-0bdc2c1d650a\" (UID: \"1b77b30e-c026-408e-a869-0bdc2c1d650a\") " Dec 06 16:06:09 crc kubenswrapper[4813]: I1206 16:06:09.761761 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b77b30e-c026-408e-a869-0bdc2c1d650a-scripts" (OuterVolumeSpecName: "scripts") pod "1b77b30e-c026-408e-a869-0bdc2c1d650a" (UID: "1b77b30e-c026-408e-a869-0bdc2c1d650a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:06:09 crc kubenswrapper[4813]: I1206 16:06:09.766217 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b77b30e-c026-408e-a869-0bdc2c1d650a-kube-api-access-qpcbk" (OuterVolumeSpecName: "kube-api-access-qpcbk") pod "1b77b30e-c026-408e-a869-0bdc2c1d650a" (UID: "1b77b30e-c026-408e-a869-0bdc2c1d650a"). InnerVolumeSpecName "kube-api-access-qpcbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:06:09 crc kubenswrapper[4813]: I1206 16:06:09.785472 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b77b30e-c026-408e-a869-0bdc2c1d650a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1b77b30e-c026-408e-a869-0bdc2c1d650a" (UID: "1b77b30e-c026-408e-a869-0bdc2c1d650a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:06:09 crc kubenswrapper[4813]: I1206 16:06:09.789190 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b77b30e-c026-408e-a869-0bdc2c1d650a-config-data" (OuterVolumeSpecName: "config-data") pod "1b77b30e-c026-408e-a869-0bdc2c1d650a" (UID: "1b77b30e-c026-408e-a869-0bdc2c1d650a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:06:09 crc kubenswrapper[4813]: I1206 16:06:09.859197 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b77b30e-c026-408e-a869-0bdc2c1d650a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:09 crc kubenswrapper[4813]: I1206 16:06:09.859227 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b77b30e-c026-408e-a869-0bdc2c1d650a-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:09 crc kubenswrapper[4813]: I1206 16:06:09.859250 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpcbk\" (UniqueName: \"kubernetes.io/projected/1b77b30e-c026-408e-a869-0bdc2c1d650a-kube-api-access-qpcbk\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:09 crc kubenswrapper[4813]: I1206 16:06:09.859267 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b77b30e-c026-408e-a869-0bdc2c1d650a-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:10 crc kubenswrapper[4813]: I1206 16:06:10.298706 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66613ee6-b0b7-48fc-adb6-86bab4940550","Type":"ContainerStarted","Data":"539e442f348f0370a3e397f4e8dc802cd0842e115a04136314b325c2298ade70"} Dec 06 16:06:10 crc kubenswrapper[4813]: I1206 16:06:10.299069 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 16:06:10 crc kubenswrapper[4813]: I1206 16:06:10.304354 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-nq29n" event={"ID":"1b77b30e-c026-408e-a869-0bdc2c1d650a","Type":"ContainerDied","Data":"4e0a8390537505e4577dc6fcbc7a9de5d385eeba6362d1c5c5a6a816c80d3cb2"} Dec 06 16:06:10 crc kubenswrapper[4813]: I1206 16:06:10.304397 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e0a8390537505e4577dc6fcbc7a9de5d385eeba6362d1c5c5a6a816c80d3cb2" Dec 06 16:06:10 crc kubenswrapper[4813]: I1206 16:06:10.304475 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-nq29n" Dec 06 16:06:10 crc kubenswrapper[4813]: I1206 16:06:10.335960 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.256097525 podStartE2EDuration="5.335938057s" podCreationTimestamp="2025-12-06 16:06:05 +0000 UTC" firstStartedPulling="2025-12-06 16:06:06.124999675 +0000 UTC m=+1206.015879261" lastFinishedPulling="2025-12-06 16:06:09.204840177 +0000 UTC m=+1209.095719793" observedRunningTime="2025-12-06 16:06:10.333719838 +0000 UTC m=+1210.224599424" watchObservedRunningTime="2025-12-06 16:06:10.335938057 +0000 UTC m=+1210.226817643" Dec 06 16:06:10 crc kubenswrapper[4813]: I1206 16:06:10.446283 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 16:06:10 crc kubenswrapper[4813]: I1206 16:06:10.446494 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ce376ede-55df-470d-bb45-584bad422181" containerName="nova-api-log" containerID="cri-o://6353afba07a44d441a2043bba5066bfa6bd9fa1ce62aa5d74d84e9969fddbf13" gracePeriod=30 Dec 06 16:06:10 crc kubenswrapper[4813]: I1206 16:06:10.446585 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ce376ede-55df-470d-bb45-584bad422181" containerName="nova-api-api" containerID="cri-o://556d98347b33cbb2eb093b02b6f4195f88bbdebb4af27a70012ab51a941682e0" gracePeriod=30 Dec 06 16:06:10 crc kubenswrapper[4813]: I1206 16:06:10.459923 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 16:06:10 crc kubenswrapper[4813]: I1206 16:06:10.460190 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="4af45f0b-4427-44d1-a612-ece893ad9b55" containerName="nova-scheduler-scheduler" containerID="cri-o://c2354e660285dee3f5e8b0c70d6303463fc1167d6c465cf4a40d704587343afc" gracePeriod=30 Dec 06 16:06:10 crc kubenswrapper[4813]: I1206 16:06:10.524916 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 16:06:10 crc kubenswrapper[4813]: I1206 16:06:10.525119 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ff13cc9e-77d0-46a8-9f98-98dd3a75cc99" containerName="nova-metadata-log" containerID="cri-o://6bb20d0d07b325fc233067fe52f03c18e26f7673230c3edb8903cb9501d81c40" gracePeriod=30 Dec 06 16:06:10 crc kubenswrapper[4813]: I1206 16:06:10.525358 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ff13cc9e-77d0-46a8-9f98-98dd3a75cc99" containerName="nova-metadata-metadata" containerID="cri-o://7facf55ea2214f87399157647a3498e61a5829cf2d0cad852b6563da23382d84" gracePeriod=30 Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.089645 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.204809 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfqzd\" (UniqueName: \"kubernetes.io/projected/ce376ede-55df-470d-bb45-584bad422181-kube-api-access-xfqzd\") pod \"ce376ede-55df-470d-bb45-584bad422181\" (UID: \"ce376ede-55df-470d-bb45-584bad422181\") " Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.205094 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce376ede-55df-470d-bb45-584bad422181-combined-ca-bundle\") pod \"ce376ede-55df-470d-bb45-584bad422181\" (UID: \"ce376ede-55df-470d-bb45-584bad422181\") " Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.205166 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce376ede-55df-470d-bb45-584bad422181-public-tls-certs\") pod \"ce376ede-55df-470d-bb45-584bad422181\" (UID: \"ce376ede-55df-470d-bb45-584bad422181\") " Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.205282 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce376ede-55df-470d-bb45-584bad422181-logs\") pod \"ce376ede-55df-470d-bb45-584bad422181\" (UID: \"ce376ede-55df-470d-bb45-584bad422181\") " Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.205425 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce376ede-55df-470d-bb45-584bad422181-internal-tls-certs\") pod \"ce376ede-55df-470d-bb45-584bad422181\" (UID: \"ce376ede-55df-470d-bb45-584bad422181\") " Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.205563 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce376ede-55df-470d-bb45-584bad422181-config-data\") pod \"ce376ede-55df-470d-bb45-584bad422181\" (UID: \"ce376ede-55df-470d-bb45-584bad422181\") " Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.220611 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce376ede-55df-470d-bb45-584bad422181-logs" (OuterVolumeSpecName: "logs") pod "ce376ede-55df-470d-bb45-584bad422181" (UID: "ce376ede-55df-470d-bb45-584bad422181"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.230273 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce376ede-55df-470d-bb45-584bad422181-kube-api-access-xfqzd" (OuterVolumeSpecName: "kube-api-access-xfqzd") pod "ce376ede-55df-470d-bb45-584bad422181" (UID: "ce376ede-55df-470d-bb45-584bad422181"). InnerVolumeSpecName "kube-api-access-xfqzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.236431 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce376ede-55df-470d-bb45-584bad422181-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ce376ede-55df-470d-bb45-584bad422181" (UID: "ce376ede-55df-470d-bb45-584bad422181"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.257128 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce376ede-55df-470d-bb45-584bad422181-config-data" (OuterVolumeSpecName: "config-data") pod "ce376ede-55df-470d-bb45-584bad422181" (UID: "ce376ede-55df-470d-bb45-584bad422181"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.308623 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfqzd\" (UniqueName: \"kubernetes.io/projected/ce376ede-55df-470d-bb45-584bad422181-kube-api-access-xfqzd\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.308646 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce376ede-55df-470d-bb45-584bad422181-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.308655 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce376ede-55df-470d-bb45-584bad422181-logs\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.308664 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce376ede-55df-470d-bb45-584bad422181-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.319418 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce376ede-55df-470d-bb45-584bad422181-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ce376ede-55df-470d-bb45-584bad422181" (UID: "ce376ede-55df-470d-bb45-584bad422181"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.344548 4813 generic.go:334] "Generic (PLEG): container finished" podID="ff13cc9e-77d0-46a8-9f98-98dd3a75cc99" containerID="6bb20d0d07b325fc233067fe52f03c18e26f7673230c3edb8903cb9501d81c40" exitCode=143 Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.344616 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99","Type":"ContainerDied","Data":"6bb20d0d07b325fc233067fe52f03c18e26f7673230c3edb8903cb9501d81c40"} Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.350371 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce376ede-55df-470d-bb45-584bad422181-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ce376ede-55df-470d-bb45-584bad422181" (UID: "ce376ede-55df-470d-bb45-584bad422181"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.352543 4813 generic.go:334] "Generic (PLEG): container finished" podID="ce376ede-55df-470d-bb45-584bad422181" containerID="556d98347b33cbb2eb093b02b6f4195f88bbdebb4af27a70012ab51a941682e0" exitCode=0 Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.352584 4813 generic.go:334] "Generic (PLEG): container finished" podID="ce376ede-55df-470d-bb45-584bad422181" containerID="6353afba07a44d441a2043bba5066bfa6bd9fa1ce62aa5d74d84e9969fddbf13" exitCode=143 Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.352670 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.352728 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce376ede-55df-470d-bb45-584bad422181","Type":"ContainerDied","Data":"556d98347b33cbb2eb093b02b6f4195f88bbdebb4af27a70012ab51a941682e0"} Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.352765 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce376ede-55df-470d-bb45-584bad422181","Type":"ContainerDied","Data":"6353afba07a44d441a2043bba5066bfa6bd9fa1ce62aa5d74d84e9969fddbf13"} Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.352779 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce376ede-55df-470d-bb45-584bad422181","Type":"ContainerDied","Data":"eac9506246601ea3f2cfe0294677c9d8a4f5c475271530791b7d676cf61ac4e0"} Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.352799 4813 scope.go:117] "RemoveContainer" containerID="556d98347b33cbb2eb093b02b6f4195f88bbdebb4af27a70012ab51a941682e0" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.413329 4813 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce376ede-55df-470d-bb45-584bad422181-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.413564 4813 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce376ede-55df-470d-bb45-584bad422181-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.419145 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.453341 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.462952 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 06 16:06:11 crc kubenswrapper[4813]: E1206 16:06:11.463425 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce376ede-55df-470d-bb45-584bad422181" containerName="nova-api-api" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.463437 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce376ede-55df-470d-bb45-584bad422181" containerName="nova-api-api" Dec 06 16:06:11 crc kubenswrapper[4813]: E1206 16:06:11.463446 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b77b30e-c026-408e-a869-0bdc2c1d650a" containerName="nova-manage" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.463452 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b77b30e-c026-408e-a869-0bdc2c1d650a" containerName="nova-manage" Dec 06 16:06:11 crc kubenswrapper[4813]: E1206 16:06:11.463478 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce376ede-55df-470d-bb45-584bad422181" containerName="nova-api-log" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.463485 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce376ede-55df-470d-bb45-584bad422181" containerName="nova-api-log" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.463676 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce376ede-55df-470d-bb45-584bad422181" containerName="nova-api-log" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.463703 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce376ede-55df-470d-bb45-584bad422181" containerName="nova-api-api" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.463719 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b77b30e-c026-408e-a869-0bdc2c1d650a" containerName="nova-manage" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.464723 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.468168 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.468193 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.468401 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.470157 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.499227 4813 scope.go:117] "RemoveContainer" containerID="6353afba07a44d441a2043bba5066bfa6bd9fa1ce62aa5d74d84e9969fddbf13" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.520789 4813 scope.go:117] "RemoveContainer" containerID="556d98347b33cbb2eb093b02b6f4195f88bbdebb4af27a70012ab51a941682e0" Dec 06 16:06:11 crc kubenswrapper[4813]: E1206 16:06:11.521493 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"556d98347b33cbb2eb093b02b6f4195f88bbdebb4af27a70012ab51a941682e0\": container with ID starting with 556d98347b33cbb2eb093b02b6f4195f88bbdebb4af27a70012ab51a941682e0 not found: ID does not exist" containerID="556d98347b33cbb2eb093b02b6f4195f88bbdebb4af27a70012ab51a941682e0" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.521538 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"556d98347b33cbb2eb093b02b6f4195f88bbdebb4af27a70012ab51a941682e0"} err="failed to get container status \"556d98347b33cbb2eb093b02b6f4195f88bbdebb4af27a70012ab51a941682e0\": rpc error: code = NotFound desc = could not find container \"556d98347b33cbb2eb093b02b6f4195f88bbdebb4af27a70012ab51a941682e0\": container with ID starting with 556d98347b33cbb2eb093b02b6f4195f88bbdebb4af27a70012ab51a941682e0 not found: ID does not exist" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.521567 4813 scope.go:117] "RemoveContainer" containerID="6353afba07a44d441a2043bba5066bfa6bd9fa1ce62aa5d74d84e9969fddbf13" Dec 06 16:06:11 crc kubenswrapper[4813]: E1206 16:06:11.526610 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6353afba07a44d441a2043bba5066bfa6bd9fa1ce62aa5d74d84e9969fddbf13\": container with ID starting with 6353afba07a44d441a2043bba5066bfa6bd9fa1ce62aa5d74d84e9969fddbf13 not found: ID does not exist" containerID="6353afba07a44d441a2043bba5066bfa6bd9fa1ce62aa5d74d84e9969fddbf13" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.526667 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6353afba07a44d441a2043bba5066bfa6bd9fa1ce62aa5d74d84e9969fddbf13"} err="failed to get container status \"6353afba07a44d441a2043bba5066bfa6bd9fa1ce62aa5d74d84e9969fddbf13\": rpc error: code = NotFound desc = could not find container \"6353afba07a44d441a2043bba5066bfa6bd9fa1ce62aa5d74d84e9969fddbf13\": container with ID starting with 6353afba07a44d441a2043bba5066bfa6bd9fa1ce62aa5d74d84e9969fddbf13 not found: ID does not exist" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.526695 4813 scope.go:117] "RemoveContainer" containerID="556d98347b33cbb2eb093b02b6f4195f88bbdebb4af27a70012ab51a941682e0" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.527043 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"556d98347b33cbb2eb093b02b6f4195f88bbdebb4af27a70012ab51a941682e0"} err="failed to get container status \"556d98347b33cbb2eb093b02b6f4195f88bbdebb4af27a70012ab51a941682e0\": rpc error: code = NotFound desc = could not find container \"556d98347b33cbb2eb093b02b6f4195f88bbdebb4af27a70012ab51a941682e0\": container with ID starting with 556d98347b33cbb2eb093b02b6f4195f88bbdebb4af27a70012ab51a941682e0 not found: ID does not exist" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.527072 4813 scope.go:117] "RemoveContainer" containerID="6353afba07a44d441a2043bba5066bfa6bd9fa1ce62aa5d74d84e9969fddbf13" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.528386 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6353afba07a44d441a2043bba5066bfa6bd9fa1ce62aa5d74d84e9969fddbf13"} err="failed to get container status \"6353afba07a44d441a2043bba5066bfa6bd9fa1ce62aa5d74d84e9969fddbf13\": rpc error: code = NotFound desc = could not find container \"6353afba07a44d441a2043bba5066bfa6bd9fa1ce62aa5d74d84e9969fddbf13\": container with ID starting with 6353afba07a44d441a2043bba5066bfa6bd9fa1ce62aa5d74d84e9969fddbf13 not found: ID does not exist" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.617925 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e54ec1f1-806c-475d-843b-53a55fea0fab-logs\") pod \"nova-api-0\" (UID: \"e54ec1f1-806c-475d-843b-53a55fea0fab\") " pod="openstack/nova-api-0" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.618007 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e54ec1f1-806c-475d-843b-53a55fea0fab-config-data\") pod \"nova-api-0\" (UID: \"e54ec1f1-806c-475d-843b-53a55fea0fab\") " pod="openstack/nova-api-0" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.618044 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e54ec1f1-806c-475d-843b-53a55fea0fab-public-tls-certs\") pod \"nova-api-0\" (UID: \"e54ec1f1-806c-475d-843b-53a55fea0fab\") " pod="openstack/nova-api-0" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.618065 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e54ec1f1-806c-475d-843b-53a55fea0fab-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e54ec1f1-806c-475d-843b-53a55fea0fab\") " pod="openstack/nova-api-0" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.618132 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xswnc\" (UniqueName: \"kubernetes.io/projected/e54ec1f1-806c-475d-843b-53a55fea0fab-kube-api-access-xswnc\") pod \"nova-api-0\" (UID: \"e54ec1f1-806c-475d-843b-53a55fea0fab\") " pod="openstack/nova-api-0" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.618174 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e54ec1f1-806c-475d-843b-53a55fea0fab-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e54ec1f1-806c-475d-843b-53a55fea0fab\") " pod="openstack/nova-api-0" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.719417 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e54ec1f1-806c-475d-843b-53a55fea0fab-logs\") pod \"nova-api-0\" (UID: \"e54ec1f1-806c-475d-843b-53a55fea0fab\") " pod="openstack/nova-api-0" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.719714 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e54ec1f1-806c-475d-843b-53a55fea0fab-config-data\") pod \"nova-api-0\" (UID: \"e54ec1f1-806c-475d-843b-53a55fea0fab\") " pod="openstack/nova-api-0" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.720438 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e54ec1f1-806c-475d-843b-53a55fea0fab-public-tls-certs\") pod \"nova-api-0\" (UID: \"e54ec1f1-806c-475d-843b-53a55fea0fab\") " pod="openstack/nova-api-0" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.720551 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e54ec1f1-806c-475d-843b-53a55fea0fab-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e54ec1f1-806c-475d-843b-53a55fea0fab\") " pod="openstack/nova-api-0" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.720640 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xswnc\" (UniqueName: \"kubernetes.io/projected/e54ec1f1-806c-475d-843b-53a55fea0fab-kube-api-access-xswnc\") pod \"nova-api-0\" (UID: \"e54ec1f1-806c-475d-843b-53a55fea0fab\") " pod="openstack/nova-api-0" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.720729 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e54ec1f1-806c-475d-843b-53a55fea0fab-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e54ec1f1-806c-475d-843b-53a55fea0fab\") " pod="openstack/nova-api-0" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.719791 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e54ec1f1-806c-475d-843b-53a55fea0fab-logs\") pod \"nova-api-0\" (UID: \"e54ec1f1-806c-475d-843b-53a55fea0fab\") " pod="openstack/nova-api-0" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.725211 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e54ec1f1-806c-475d-843b-53a55fea0fab-public-tls-certs\") pod \"nova-api-0\" (UID: \"e54ec1f1-806c-475d-843b-53a55fea0fab\") " pod="openstack/nova-api-0" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.725248 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e54ec1f1-806c-475d-843b-53a55fea0fab-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e54ec1f1-806c-475d-843b-53a55fea0fab\") " pod="openstack/nova-api-0" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.726439 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e54ec1f1-806c-475d-843b-53a55fea0fab-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e54ec1f1-806c-475d-843b-53a55fea0fab\") " pod="openstack/nova-api-0" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.727721 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e54ec1f1-806c-475d-843b-53a55fea0fab-config-data\") pod \"nova-api-0\" (UID: \"e54ec1f1-806c-475d-843b-53a55fea0fab\") " pod="openstack/nova-api-0" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.743161 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xswnc\" (UniqueName: \"kubernetes.io/projected/e54ec1f1-806c-475d-843b-53a55fea0fab-kube-api-access-xswnc\") pod \"nova-api-0\" (UID: \"e54ec1f1-806c-475d-843b-53a55fea0fab\") " pod="openstack/nova-api-0" Dec 06 16:06:11 crc kubenswrapper[4813]: I1206 16:06:11.827940 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 16:06:12 crc kubenswrapper[4813]: I1206 16:06:12.246947 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 16:06:12 crc kubenswrapper[4813]: I1206 16:06:12.384505 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e54ec1f1-806c-475d-843b-53a55fea0fab","Type":"ContainerStarted","Data":"2649a7e21ccbe3ef44ed45e00a2fb471440cc05fe5ee8e5fdbcc7fa529972354"} Dec 06 16:06:12 crc kubenswrapper[4813]: I1206 16:06:12.496357 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce376ede-55df-470d-bb45-584bad422181" path="/var/lib/kubelet/pods/ce376ede-55df-470d-bb45-584bad422181/volumes" Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.370344 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.390114 4813 generic.go:334] "Generic (PLEG): container finished" podID="4af45f0b-4427-44d1-a612-ece893ad9b55" containerID="c2354e660285dee3f5e8b0c70d6303463fc1167d6c465cf4a40d704587343afc" exitCode=0 Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.390158 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4af45f0b-4427-44d1-a612-ece893ad9b55","Type":"ContainerDied","Data":"c2354e660285dee3f5e8b0c70d6303463fc1167d6c465cf4a40d704587343afc"} Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.390182 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4af45f0b-4427-44d1-a612-ece893ad9b55","Type":"ContainerDied","Data":"9f3e2b1953cea6dcacc64a91ace828be60e134a35d4e91b08fac0dde15ae985d"} Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.390197 4813 scope.go:117] "RemoveContainer" containerID="c2354e660285dee3f5e8b0c70d6303463fc1167d6c465cf4a40d704587343afc" Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.390281 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.392675 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e54ec1f1-806c-475d-843b-53a55fea0fab","Type":"ContainerStarted","Data":"fbe9e383252683615a6c9a81268d0d68815a29f060ad722bb73e86b1a5d8dd5d"} Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.392697 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e54ec1f1-806c-475d-843b-53a55fea0fab","Type":"ContainerStarted","Data":"af3950f152a51b72ec32375549e0cdac68540f2d9bfbb8f1ade6884cd268a965"} Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.409860 4813 scope.go:117] "RemoveContainer" containerID="c2354e660285dee3f5e8b0c70d6303463fc1167d6c465cf4a40d704587343afc" Dec 06 16:06:13 crc kubenswrapper[4813]: E1206 16:06:13.410210 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2354e660285dee3f5e8b0c70d6303463fc1167d6c465cf4a40d704587343afc\": container with ID starting with c2354e660285dee3f5e8b0c70d6303463fc1167d6c465cf4a40d704587343afc not found: ID does not exist" containerID="c2354e660285dee3f5e8b0c70d6303463fc1167d6c465cf4a40d704587343afc" Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.410244 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2354e660285dee3f5e8b0c70d6303463fc1167d6c465cf4a40d704587343afc"} err="failed to get container status \"c2354e660285dee3f5e8b0c70d6303463fc1167d6c465cf4a40d704587343afc\": rpc error: code = NotFound desc = could not find container \"c2354e660285dee3f5e8b0c70d6303463fc1167d6c465cf4a40d704587343afc\": container with ID starting with c2354e660285dee3f5e8b0c70d6303463fc1167d6c465cf4a40d704587343afc not found: ID does not exist" Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.436705 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.436687374 podStartE2EDuration="2.436687374s" podCreationTimestamp="2025-12-06 16:06:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:06:13.43578707 +0000 UTC m=+1213.326666646" watchObservedRunningTime="2025-12-06 16:06:13.436687374 +0000 UTC m=+1213.327566950" Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.459454 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af45f0b-4427-44d1-a612-ece893ad9b55-combined-ca-bundle\") pod \"4af45f0b-4427-44d1-a612-ece893ad9b55\" (UID: \"4af45f0b-4427-44d1-a612-ece893ad9b55\") " Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.459527 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5clb\" (UniqueName: \"kubernetes.io/projected/4af45f0b-4427-44d1-a612-ece893ad9b55-kube-api-access-z5clb\") pod \"4af45f0b-4427-44d1-a612-ece893ad9b55\" (UID: \"4af45f0b-4427-44d1-a612-ece893ad9b55\") " Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.459609 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4af45f0b-4427-44d1-a612-ece893ad9b55-config-data\") pod \"4af45f0b-4427-44d1-a612-ece893ad9b55\" (UID: \"4af45f0b-4427-44d1-a612-ece893ad9b55\") " Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.465566 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4af45f0b-4427-44d1-a612-ece893ad9b55-kube-api-access-z5clb" (OuterVolumeSpecName: "kube-api-access-z5clb") pod "4af45f0b-4427-44d1-a612-ece893ad9b55" (UID: "4af45f0b-4427-44d1-a612-ece893ad9b55"). InnerVolumeSpecName "kube-api-access-z5clb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.490893 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4af45f0b-4427-44d1-a612-ece893ad9b55-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4af45f0b-4427-44d1-a612-ece893ad9b55" (UID: "4af45f0b-4427-44d1-a612-ece893ad9b55"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.513138 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4af45f0b-4427-44d1-a612-ece893ad9b55-config-data" (OuterVolumeSpecName: "config-data") pod "4af45f0b-4427-44d1-a612-ece893ad9b55" (UID: "4af45f0b-4427-44d1-a612-ece893ad9b55"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.561386 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af45f0b-4427-44d1-a612-ece893ad9b55-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.561426 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5clb\" (UniqueName: \"kubernetes.io/projected/4af45f0b-4427-44d1-a612-ece893ad9b55-kube-api-access-z5clb\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.561446 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4af45f0b-4427-44d1-a612-ece893ad9b55-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.720401 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.726941 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.743465 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 16:06:13 crc kubenswrapper[4813]: E1206 16:06:13.743870 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4af45f0b-4427-44d1-a612-ece893ad9b55" containerName="nova-scheduler-scheduler" Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.743892 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="4af45f0b-4427-44d1-a612-ece893ad9b55" containerName="nova-scheduler-scheduler" Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.744087 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="4af45f0b-4427-44d1-a612-ece893ad9b55" containerName="nova-scheduler-scheduler" Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.744675 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.747810 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.773229 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.866312 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d53150b5-8980-47ff-bc3a-6b75ac75fa14-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d53150b5-8980-47ff-bc3a-6b75ac75fa14\") " pod="openstack/nova-scheduler-0" Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.866587 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsklq\" (UniqueName: \"kubernetes.io/projected/d53150b5-8980-47ff-bc3a-6b75ac75fa14-kube-api-access-vsklq\") pod \"nova-scheduler-0\" (UID: \"d53150b5-8980-47ff-bc3a-6b75ac75fa14\") " pod="openstack/nova-scheduler-0" Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.866625 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d53150b5-8980-47ff-bc3a-6b75ac75fa14-config-data\") pod \"nova-scheduler-0\" (UID: \"d53150b5-8980-47ff-bc3a-6b75ac75fa14\") " pod="openstack/nova-scheduler-0" Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.952769 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="ff13cc9e-77d0-46a8-9f98-98dd3a75cc99" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": read tcp 10.217.0.2:41844->10.217.0.193:8775: read: connection reset by peer" Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.953084 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="ff13cc9e-77d0-46a8-9f98-98dd3a75cc99" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": read tcp 10.217.0.2:41828->10.217.0.193:8775: read: connection reset by peer" Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.967796 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsklq\" (UniqueName: \"kubernetes.io/projected/d53150b5-8980-47ff-bc3a-6b75ac75fa14-kube-api-access-vsklq\") pod \"nova-scheduler-0\" (UID: \"d53150b5-8980-47ff-bc3a-6b75ac75fa14\") " pod="openstack/nova-scheduler-0" Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.968037 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d53150b5-8980-47ff-bc3a-6b75ac75fa14-config-data\") pod \"nova-scheduler-0\" (UID: \"d53150b5-8980-47ff-bc3a-6b75ac75fa14\") " pod="openstack/nova-scheduler-0" Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.968305 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d53150b5-8980-47ff-bc3a-6b75ac75fa14-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d53150b5-8980-47ff-bc3a-6b75ac75fa14\") " pod="openstack/nova-scheduler-0" Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.972789 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d53150b5-8980-47ff-bc3a-6b75ac75fa14-config-data\") pod \"nova-scheduler-0\" (UID: \"d53150b5-8980-47ff-bc3a-6b75ac75fa14\") " pod="openstack/nova-scheduler-0" Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.973713 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d53150b5-8980-47ff-bc3a-6b75ac75fa14-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d53150b5-8980-47ff-bc3a-6b75ac75fa14\") " pod="openstack/nova-scheduler-0" Dec 06 16:06:13 crc kubenswrapper[4813]: I1206 16:06:13.992775 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsklq\" (UniqueName: \"kubernetes.io/projected/d53150b5-8980-47ff-bc3a-6b75ac75fa14-kube-api-access-vsklq\") pod \"nova-scheduler-0\" (UID: \"d53150b5-8980-47ff-bc3a-6b75ac75fa14\") " pod="openstack/nova-scheduler-0" Dec 06 16:06:14 crc kubenswrapper[4813]: I1206 16:06:14.058399 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 16:06:14 crc kubenswrapper[4813]: I1206 16:06:14.437396 4813 generic.go:334] "Generic (PLEG): container finished" podID="ff13cc9e-77d0-46a8-9f98-98dd3a75cc99" containerID="7facf55ea2214f87399157647a3498e61a5829cf2d0cad852b6563da23382d84" exitCode=0 Dec 06 16:06:14 crc kubenswrapper[4813]: I1206 16:06:14.438371 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99","Type":"ContainerDied","Data":"7facf55ea2214f87399157647a3498e61a5829cf2d0cad852b6563da23382d84"} Dec 06 16:06:14 crc kubenswrapper[4813]: I1206 16:06:14.510363 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4af45f0b-4427-44d1-a612-ece893ad9b55" path="/var/lib/kubelet/pods/4af45f0b-4427-44d1-a612-ece893ad9b55/volumes" Dec 06 16:06:14 crc kubenswrapper[4813]: I1206 16:06:14.651475 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 16:06:14 crc kubenswrapper[4813]: I1206 16:06:14.736746 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 16:06:14 crc kubenswrapper[4813]: I1206 16:06:14.788054 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99-nova-metadata-tls-certs\") pod \"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99\" (UID: \"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99\") " Dec 06 16:06:14 crc kubenswrapper[4813]: I1206 16:06:14.788373 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfqfk\" (UniqueName: \"kubernetes.io/projected/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99-kube-api-access-jfqfk\") pod \"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99\" (UID: \"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99\") " Dec 06 16:06:14 crc kubenswrapper[4813]: I1206 16:06:14.788481 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99-logs\") pod \"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99\" (UID: \"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99\") " Dec 06 16:06:14 crc kubenswrapper[4813]: I1206 16:06:14.788529 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99-combined-ca-bundle\") pod \"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99\" (UID: \"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99\") " Dec 06 16:06:14 crc kubenswrapper[4813]: I1206 16:06:14.788580 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99-config-data\") pod \"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99\" (UID: \"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99\") " Dec 06 16:06:14 crc kubenswrapper[4813]: I1206 16:06:14.789783 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99-logs" (OuterVolumeSpecName: "logs") pod "ff13cc9e-77d0-46a8-9f98-98dd3a75cc99" (UID: "ff13cc9e-77d0-46a8-9f98-98dd3a75cc99"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:06:14 crc kubenswrapper[4813]: I1206 16:06:14.801380 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99-kube-api-access-jfqfk" (OuterVolumeSpecName: "kube-api-access-jfqfk") pod "ff13cc9e-77d0-46a8-9f98-98dd3a75cc99" (UID: "ff13cc9e-77d0-46a8-9f98-98dd3a75cc99"). InnerVolumeSpecName "kube-api-access-jfqfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:06:14 crc kubenswrapper[4813]: I1206 16:06:14.828814 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff13cc9e-77d0-46a8-9f98-98dd3a75cc99" (UID: "ff13cc9e-77d0-46a8-9f98-98dd3a75cc99"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:06:14 crc kubenswrapper[4813]: I1206 16:06:14.832379 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99-config-data" (OuterVolumeSpecName: "config-data") pod "ff13cc9e-77d0-46a8-9f98-98dd3a75cc99" (UID: "ff13cc9e-77d0-46a8-9f98-98dd3a75cc99"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:06:14 crc kubenswrapper[4813]: I1206 16:06:14.870446 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "ff13cc9e-77d0-46a8-9f98-98dd3a75cc99" (UID: "ff13cc9e-77d0-46a8-9f98-98dd3a75cc99"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:06:14 crc kubenswrapper[4813]: I1206 16:06:14.890756 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:14 crc kubenswrapper[4813]: I1206 16:06:14.890786 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:14 crc kubenswrapper[4813]: I1206 16:06:14.890796 4813 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:14 crc kubenswrapper[4813]: I1206 16:06:14.890824 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jfqfk\" (UniqueName: \"kubernetes.io/projected/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99-kube-api-access-jfqfk\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:14 crc kubenswrapper[4813]: I1206 16:06:14.890834 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99-logs\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.448381 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d53150b5-8980-47ff-bc3a-6b75ac75fa14","Type":"ContainerStarted","Data":"369fa66b9172d576e638aeee6bc9f3af0bb6b40fa720c3a3c1bb2a1ad736a5a1"} Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.448675 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d53150b5-8980-47ff-bc3a-6b75ac75fa14","Type":"ContainerStarted","Data":"8d80b2920863ea4a9d20918199be5de08f3c0fc0760c58aafbd402b13689278f"} Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.450695 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ff13cc9e-77d0-46a8-9f98-98dd3a75cc99","Type":"ContainerDied","Data":"74933231fba764bb2aaf2412fb2ff42d397750898954c85ce628debcd914ec63"} Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.450744 4813 scope.go:117] "RemoveContainer" containerID="7facf55ea2214f87399157647a3498e61a5829cf2d0cad852b6563da23382d84" Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.450774 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.471619 4813 scope.go:117] "RemoveContainer" containerID="6bb20d0d07b325fc233067fe52f03c18e26f7673230c3edb8903cb9501d81c40" Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.492301 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.49223882 podStartE2EDuration="2.49223882s" podCreationTimestamp="2025-12-06 16:06:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:06:15.465181364 +0000 UTC m=+1215.356060940" watchObservedRunningTime="2025-12-06 16:06:15.49223882 +0000 UTC m=+1215.383118426" Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.534010 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.569287 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.582310 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 06 16:06:15 crc kubenswrapper[4813]: E1206 16:06:15.582733 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff13cc9e-77d0-46a8-9f98-98dd3a75cc99" containerName="nova-metadata-metadata" Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.582752 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff13cc9e-77d0-46a8-9f98-98dd3a75cc99" containerName="nova-metadata-metadata" Dec 06 16:06:15 crc kubenswrapper[4813]: E1206 16:06:15.582780 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff13cc9e-77d0-46a8-9f98-98dd3a75cc99" containerName="nova-metadata-log" Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.582787 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff13cc9e-77d0-46a8-9f98-98dd3a75cc99" containerName="nova-metadata-log" Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.582977 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff13cc9e-77d0-46a8-9f98-98dd3a75cc99" containerName="nova-metadata-metadata" Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.582996 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff13cc9e-77d0-46a8-9f98-98dd3a75cc99" containerName="nova-metadata-log" Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.583965 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.586921 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.587734 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.594510 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.738624 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1731c7e7-8ded-4a61-8b8a-d94cf8b4a860-logs\") pod \"nova-metadata-0\" (UID: \"1731c7e7-8ded-4a61-8b8a-d94cf8b4a860\") " pod="openstack/nova-metadata-0" Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.738677 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1731c7e7-8ded-4a61-8b8a-d94cf8b4a860-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1731c7e7-8ded-4a61-8b8a-d94cf8b4a860\") " pod="openstack/nova-metadata-0" Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.738729 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1731c7e7-8ded-4a61-8b8a-d94cf8b4a860-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1731c7e7-8ded-4a61-8b8a-d94cf8b4a860\") " pod="openstack/nova-metadata-0" Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.738764 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1731c7e7-8ded-4a61-8b8a-d94cf8b4a860-config-data\") pod \"nova-metadata-0\" (UID: \"1731c7e7-8ded-4a61-8b8a-d94cf8b4a860\") " pod="openstack/nova-metadata-0" Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.738797 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stzzg\" (UniqueName: \"kubernetes.io/projected/1731c7e7-8ded-4a61-8b8a-d94cf8b4a860-kube-api-access-stzzg\") pod \"nova-metadata-0\" (UID: \"1731c7e7-8ded-4a61-8b8a-d94cf8b4a860\") " pod="openstack/nova-metadata-0" Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.840067 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1731c7e7-8ded-4a61-8b8a-d94cf8b4a860-logs\") pod \"nova-metadata-0\" (UID: \"1731c7e7-8ded-4a61-8b8a-d94cf8b4a860\") " pod="openstack/nova-metadata-0" Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.840124 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1731c7e7-8ded-4a61-8b8a-d94cf8b4a860-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1731c7e7-8ded-4a61-8b8a-d94cf8b4a860\") " pod="openstack/nova-metadata-0" Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.840188 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1731c7e7-8ded-4a61-8b8a-d94cf8b4a860-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1731c7e7-8ded-4a61-8b8a-d94cf8b4a860\") " pod="openstack/nova-metadata-0" Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.840225 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1731c7e7-8ded-4a61-8b8a-d94cf8b4a860-config-data\") pod \"nova-metadata-0\" (UID: \"1731c7e7-8ded-4a61-8b8a-d94cf8b4a860\") " pod="openstack/nova-metadata-0" Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.840276 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stzzg\" (UniqueName: \"kubernetes.io/projected/1731c7e7-8ded-4a61-8b8a-d94cf8b4a860-kube-api-access-stzzg\") pod \"nova-metadata-0\" (UID: \"1731c7e7-8ded-4a61-8b8a-d94cf8b4a860\") " pod="openstack/nova-metadata-0" Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.840866 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1731c7e7-8ded-4a61-8b8a-d94cf8b4a860-logs\") pod \"nova-metadata-0\" (UID: \"1731c7e7-8ded-4a61-8b8a-d94cf8b4a860\") " pod="openstack/nova-metadata-0" Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.846983 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1731c7e7-8ded-4a61-8b8a-d94cf8b4a860-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1731c7e7-8ded-4a61-8b8a-d94cf8b4a860\") " pod="openstack/nova-metadata-0" Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.847104 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1731c7e7-8ded-4a61-8b8a-d94cf8b4a860-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1731c7e7-8ded-4a61-8b8a-d94cf8b4a860\") " pod="openstack/nova-metadata-0" Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.847717 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1731c7e7-8ded-4a61-8b8a-d94cf8b4a860-config-data\") pod \"nova-metadata-0\" (UID: \"1731c7e7-8ded-4a61-8b8a-d94cf8b4a860\") " pod="openstack/nova-metadata-0" Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.858775 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stzzg\" (UniqueName: \"kubernetes.io/projected/1731c7e7-8ded-4a61-8b8a-d94cf8b4a860-kube-api-access-stzzg\") pod \"nova-metadata-0\" (UID: \"1731c7e7-8ded-4a61-8b8a-d94cf8b4a860\") " pod="openstack/nova-metadata-0" Dec 06 16:06:15 crc kubenswrapper[4813]: I1206 16:06:15.912520 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 16:06:16 crc kubenswrapper[4813]: I1206 16:06:16.375134 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 16:06:16 crc kubenswrapper[4813]: I1206 16:06:16.462908 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1731c7e7-8ded-4a61-8b8a-d94cf8b4a860","Type":"ContainerStarted","Data":"492acc8910f5ee9d9045464c341d844bc97c0069160bcf68b0d5ad203ba19223"} Dec 06 16:06:16 crc kubenswrapper[4813]: I1206 16:06:16.504139 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff13cc9e-77d0-46a8-9f98-98dd3a75cc99" path="/var/lib/kubelet/pods/ff13cc9e-77d0-46a8-9f98-98dd3a75cc99/volumes" Dec 06 16:06:17 crc kubenswrapper[4813]: I1206 16:06:17.475204 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1731c7e7-8ded-4a61-8b8a-d94cf8b4a860","Type":"ContainerStarted","Data":"7aea97f2d02f213e375d7aad31c47d5428c6d27c20b5a8cfda262e1e0df1c5e8"} Dec 06 16:06:17 crc kubenswrapper[4813]: I1206 16:06:17.476329 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1731c7e7-8ded-4a61-8b8a-d94cf8b4a860","Type":"ContainerStarted","Data":"8e23729cba04bbbfbcbb39378abc8ce3cbfbf8561d74b201e8a5147d7a81ebbb"} Dec 06 16:06:17 crc kubenswrapper[4813]: I1206 16:06:17.501430 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.50141092 podStartE2EDuration="2.50141092s" podCreationTimestamp="2025-12-06 16:06:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:06:17.499040357 +0000 UTC m=+1217.389919983" watchObservedRunningTime="2025-12-06 16:06:17.50141092 +0000 UTC m=+1217.392290486" Dec 06 16:06:19 crc kubenswrapper[4813]: I1206 16:06:19.058677 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 06 16:06:19 crc kubenswrapper[4813]: I1206 16:06:19.427662 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:06:19 crc kubenswrapper[4813]: I1206 16:06:19.427755 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:06:19 crc kubenswrapper[4813]: I1206 16:06:19.427831 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" Dec 06 16:06:19 crc kubenswrapper[4813]: I1206 16:06:19.428704 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b7b68c16444a3326ba33662bc5c9583c0b689e2b31a8a7643be0512e87a18c96"} pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 16:06:19 crc kubenswrapper[4813]: I1206 16:06:19.428812 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" containerID="cri-o://b7b68c16444a3326ba33662bc5c9583c0b689e2b31a8a7643be0512e87a18c96" gracePeriod=600 Dec 06 16:06:20 crc kubenswrapper[4813]: I1206 16:06:20.508480 4813 generic.go:334] "Generic (PLEG): container finished" podID="d88e8bae-c055-4c55-b548-f621ff96de06" containerID="b7b68c16444a3326ba33662bc5c9583c0b689e2b31a8a7643be0512e87a18c96" exitCode=0 Dec 06 16:06:20 crc kubenswrapper[4813]: I1206 16:06:20.508803 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerDied","Data":"b7b68c16444a3326ba33662bc5c9583c0b689e2b31a8a7643be0512e87a18c96"} Dec 06 16:06:20 crc kubenswrapper[4813]: I1206 16:06:20.508834 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerStarted","Data":"f25008e92574e2c832578c1ff4d4518d3c2eb82b415ff9e3dd99f9504d835fc6"} Dec 06 16:06:20 crc kubenswrapper[4813]: I1206 16:06:20.508853 4813 scope.go:117] "RemoveContainer" containerID="659cf1c13b1c009e3dc025bbc1b70aff2561005d40bd452ea21f924d8a1cca89" Dec 06 16:06:20 crc kubenswrapper[4813]: I1206 16:06:20.913018 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 16:06:20 crc kubenswrapper[4813]: I1206 16:06:20.913099 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 16:06:21 crc kubenswrapper[4813]: I1206 16:06:21.829174 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 16:06:21 crc kubenswrapper[4813]: I1206 16:06:21.829855 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 16:06:22 crc kubenswrapper[4813]: I1206 16:06:22.846412 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e54ec1f1-806c-475d-843b-53a55fea0fab" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 16:06:22 crc kubenswrapper[4813]: I1206 16:06:22.846412 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e54ec1f1-806c-475d-843b-53a55fea0fab" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 16:06:24 crc kubenswrapper[4813]: I1206 16:06:24.058583 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 06 16:06:24 crc kubenswrapper[4813]: I1206 16:06:24.103528 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 06 16:06:24 crc kubenswrapper[4813]: I1206 16:06:24.608778 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 06 16:06:25 crc kubenswrapper[4813]: I1206 16:06:25.913474 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 16:06:25 crc kubenswrapper[4813]: I1206 16:06:25.913761 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 16:06:26 crc kubenswrapper[4813]: I1206 16:06:26.933436 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1731c7e7-8ded-4a61-8b8a-d94cf8b4a860" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.206:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 16:06:26 crc kubenswrapper[4813]: I1206 16:06:26.933490 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1731c7e7-8ded-4a61-8b8a-d94cf8b4a860" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.206:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 16:06:31 crc kubenswrapper[4813]: I1206 16:06:31.834051 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 06 16:06:31 crc kubenswrapper[4813]: I1206 16:06:31.834393 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 06 16:06:31 crc kubenswrapper[4813]: I1206 16:06:31.834729 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 06 16:06:31 crc kubenswrapper[4813]: I1206 16:06:31.835229 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 06 16:06:31 crc kubenswrapper[4813]: I1206 16:06:31.843585 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 06 16:06:31 crc kubenswrapper[4813]: I1206 16:06:31.845234 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 06 16:06:35 crc kubenswrapper[4813]: I1206 16:06:35.652576 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 06 16:06:35 crc kubenswrapper[4813]: I1206 16:06:35.920429 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 06 16:06:35 crc kubenswrapper[4813]: I1206 16:06:35.923669 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 06 16:06:35 crc kubenswrapper[4813]: I1206 16:06:35.936833 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 06 16:06:36 crc kubenswrapper[4813]: I1206 16:06:36.690903 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 06 16:06:45 crc kubenswrapper[4813]: I1206 16:06:45.677034 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 16:06:46 crc kubenswrapper[4813]: I1206 16:06:46.994213 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 16:06:50 crc kubenswrapper[4813]: I1206 16:06:50.000073 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="f03c3251-5e1c-4cb2-aa3f-ebbef28acb19" containerName="rabbitmq" containerID="cri-o://8e69c41db3d3d7e419c2bdde0dabb053b2cf3846832653472e595e340e23a8a7" gracePeriod=604796 Dec 06 16:06:50 crc kubenswrapper[4813]: I1206 16:06:50.687309 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="f03c3251-5e1c-4cb2-aa3f-ebbef28acb19" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Dec 06 16:06:51 crc kubenswrapper[4813]: I1206 16:06:51.204785 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="2172dfb5-c0a6-4525-a326-dd9736eb5233" containerName="rabbitmq" containerID="cri-o://43a1c843efdf2356e06677d802b023e646e54b4c679dfa4d791704c633a487bd" gracePeriod=604796 Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.725948 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.838257 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.838385 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-server-conf\") pod \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.838446 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzdk5\" (UniqueName: \"kubernetes.io/projected/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-kube-api-access-rzdk5\") pod \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.838511 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-erlang-cookie-secret\") pod \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.838593 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-rabbitmq-plugins\") pod \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.838662 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-rabbitmq-confd\") pod \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.838718 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-pod-info\") pod \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.838771 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-config-data\") pod \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.838812 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-plugins-conf\") pod \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.838880 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-rabbitmq-erlang-cookie\") pod \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.838910 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-rabbitmq-tls\") pod \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\" (UID: \"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19\") " Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.847759 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "f03c3251-5e1c-4cb2-aa3f-ebbef28acb19" (UID: "f03c3251-5e1c-4cb2-aa3f-ebbef28acb19"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.853799 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-pod-info" (OuterVolumeSpecName: "pod-info") pod "f03c3251-5e1c-4cb2-aa3f-ebbef28acb19" (UID: "f03c3251-5e1c-4cb2-aa3f-ebbef28acb19"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.854185 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-kube-api-access-rzdk5" (OuterVolumeSpecName: "kube-api-access-rzdk5") pod "f03c3251-5e1c-4cb2-aa3f-ebbef28acb19" (UID: "f03c3251-5e1c-4cb2-aa3f-ebbef28acb19"). InnerVolumeSpecName "kube-api-access-rzdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.854683 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "f03c3251-5e1c-4cb2-aa3f-ebbef28acb19" (UID: "f03c3251-5e1c-4cb2-aa3f-ebbef28acb19"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.854829 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "f03c3251-5e1c-4cb2-aa3f-ebbef28acb19" (UID: "f03c3251-5e1c-4cb2-aa3f-ebbef28acb19"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.858359 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "f03c3251-5e1c-4cb2-aa3f-ebbef28acb19" (UID: "f03c3251-5e1c-4cb2-aa3f-ebbef28acb19"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.859417 4813 generic.go:334] "Generic (PLEG): container finished" podID="f03c3251-5e1c-4cb2-aa3f-ebbef28acb19" containerID="8e69c41db3d3d7e419c2bdde0dabb053b2cf3846832653472e595e340e23a8a7" exitCode=0 Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.859456 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19","Type":"ContainerDied","Data":"8e69c41db3d3d7e419c2bdde0dabb053b2cf3846832653472e595e340e23a8a7"} Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.859506 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f03c3251-5e1c-4cb2-aa3f-ebbef28acb19","Type":"ContainerDied","Data":"a58091e3567d12b58c4bbae8d972435c43fedb3d6e70bdf69419dba603eaedea"} Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.859524 4813 scope.go:117] "RemoveContainer" containerID="8e69c41db3d3d7e419c2bdde0dabb053b2cf3846832653472e595e340e23a8a7" Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.859646 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.874058 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "f03c3251-5e1c-4cb2-aa3f-ebbef28acb19" (UID: "f03c3251-5e1c-4cb2-aa3f-ebbef28acb19"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.875394 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "f03c3251-5e1c-4cb2-aa3f-ebbef28acb19" (UID: "f03c3251-5e1c-4cb2-aa3f-ebbef28acb19"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.891536 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-config-data" (OuterVolumeSpecName: "config-data") pod "f03c3251-5e1c-4cb2-aa3f-ebbef28acb19" (UID: "f03c3251-5e1c-4cb2-aa3f-ebbef28acb19"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.894917 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-server-conf" (OuterVolumeSpecName: "server-conf") pod "f03c3251-5e1c-4cb2-aa3f-ebbef28acb19" (UID: "f03c3251-5e1c-4cb2-aa3f-ebbef28acb19"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.940700 4813 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-pod-info\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.940866 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.940934 4813 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.940997 4813 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.941571 4813 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.941658 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.941720 4813 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-server-conf\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.941780 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rzdk5\" (UniqueName: \"kubernetes.io/projected/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-kube-api-access-rzdk5\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.941835 4813 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.941914 4813 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.945789 4813 scope.go:117] "RemoveContainer" containerID="d2852b625a8e7a7b46550b6c2cafd975948130b8c072519d3613885a39ad6c8f" Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.976510 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "f03c3251-5e1c-4cb2-aa3f-ebbef28acb19" (UID: "f03c3251-5e1c-4cb2-aa3f-ebbef28acb19"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.997609 4813 scope.go:117] "RemoveContainer" containerID="8e69c41db3d3d7e419c2bdde0dabb053b2cf3846832653472e595e340e23a8a7" Dec 06 16:06:56 crc kubenswrapper[4813]: E1206 16:06:56.998798 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e69c41db3d3d7e419c2bdde0dabb053b2cf3846832653472e595e340e23a8a7\": container with ID starting with 8e69c41db3d3d7e419c2bdde0dabb053b2cf3846832653472e595e340e23a8a7 not found: ID does not exist" containerID="8e69c41db3d3d7e419c2bdde0dabb053b2cf3846832653472e595e340e23a8a7" Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.998834 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e69c41db3d3d7e419c2bdde0dabb053b2cf3846832653472e595e340e23a8a7"} err="failed to get container status \"8e69c41db3d3d7e419c2bdde0dabb053b2cf3846832653472e595e340e23a8a7\": rpc error: code = NotFound desc = could not find container \"8e69c41db3d3d7e419c2bdde0dabb053b2cf3846832653472e595e340e23a8a7\": container with ID starting with 8e69c41db3d3d7e419c2bdde0dabb053b2cf3846832653472e595e340e23a8a7 not found: ID does not exist" Dec 06 16:06:56 crc kubenswrapper[4813]: I1206 16:06:56.998873 4813 scope.go:117] "RemoveContainer" containerID="d2852b625a8e7a7b46550b6c2cafd975948130b8c072519d3613885a39ad6c8f" Dec 06 16:06:57 crc kubenswrapper[4813]: E1206 16:06:57.001805 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2852b625a8e7a7b46550b6c2cafd975948130b8c072519d3613885a39ad6c8f\": container with ID starting with d2852b625a8e7a7b46550b6c2cafd975948130b8c072519d3613885a39ad6c8f not found: ID does not exist" containerID="d2852b625a8e7a7b46550b6c2cafd975948130b8c072519d3613885a39ad6c8f" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.001844 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2852b625a8e7a7b46550b6c2cafd975948130b8c072519d3613885a39ad6c8f"} err="failed to get container status \"d2852b625a8e7a7b46550b6c2cafd975948130b8c072519d3613885a39ad6c8f\": rpc error: code = NotFound desc = could not find container \"d2852b625a8e7a7b46550b6c2cafd975948130b8c072519d3613885a39ad6c8f\": container with ID starting with d2852b625a8e7a7b46550b6c2cafd975948130b8c072519d3613885a39ad6c8f not found: ID does not exist" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.004417 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.043568 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.043609 4813 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.190174 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.198331 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.221283 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 16:06:57 crc kubenswrapper[4813]: E1206 16:06:57.221664 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f03c3251-5e1c-4cb2-aa3f-ebbef28acb19" containerName="rabbitmq" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.221682 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f03c3251-5e1c-4cb2-aa3f-ebbef28acb19" containerName="rabbitmq" Dec 06 16:06:57 crc kubenswrapper[4813]: E1206 16:06:57.221708 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f03c3251-5e1c-4cb2-aa3f-ebbef28acb19" containerName="setup-container" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.221714 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f03c3251-5e1c-4cb2-aa3f-ebbef28acb19" containerName="setup-container" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.221920 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f03c3251-5e1c-4cb2-aa3f-ebbef28acb19" containerName="rabbitmq" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.222879 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.225603 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.226364 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.229813 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.230361 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.230364 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.230558 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.230584 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-nndc8" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.247469 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5ef242c1-0127-4963-9563-4943a584398c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.247545 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5ef242c1-0127-4963-9563-4943a584398c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.247640 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5ef242c1-0127-4963-9563-4943a584398c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.247736 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.247848 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5ef242c1-0127-4963-9563-4943a584398c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.247908 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5ef242c1-0127-4963-9563-4943a584398c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.247960 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5ef242c1-0127-4963-9563-4943a584398c-config-data\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.248030 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5ef242c1-0127-4963-9563-4943a584398c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.248127 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5ef242c1-0127-4963-9563-4943a584398c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.248273 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5ef242c1-0127-4963-9563-4943a584398c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.248297 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vmj7\" (UniqueName: \"kubernetes.io/projected/5ef242c1-0127-4963-9563-4943a584398c-kube-api-access-7vmj7\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.258367 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.349687 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5ef242c1-0127-4963-9563-4943a584398c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.349735 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5ef242c1-0127-4963-9563-4943a584398c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.350127 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5ef242c1-0127-4963-9563-4943a584398c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.350197 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.350310 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5ef242c1-0127-4963-9563-4943a584398c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.350349 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5ef242c1-0127-4963-9563-4943a584398c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.350384 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5ef242c1-0127-4963-9563-4943a584398c-config-data\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.350421 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5ef242c1-0127-4963-9563-4943a584398c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.350448 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5ef242c1-0127-4963-9563-4943a584398c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.350499 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5ef242c1-0127-4963-9563-4943a584398c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.350517 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vmj7\" (UniqueName: \"kubernetes.io/projected/5ef242c1-0127-4963-9563-4943a584398c-kube-api-access-7vmj7\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.350996 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5ef242c1-0127-4963-9563-4943a584398c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.350495 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.351304 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5ef242c1-0127-4963-9563-4943a584398c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.352316 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5ef242c1-0127-4963-9563-4943a584398c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.352459 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5ef242c1-0127-4963-9563-4943a584398c-config-data\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.353493 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5ef242c1-0127-4963-9563-4943a584398c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.356364 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5ef242c1-0127-4963-9563-4943a584398c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.357814 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5ef242c1-0127-4963-9563-4943a584398c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.362903 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5ef242c1-0127-4963-9563-4943a584398c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.363374 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5ef242c1-0127-4963-9563-4943a584398c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.370382 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vmj7\" (UniqueName: \"kubernetes.io/projected/5ef242c1-0127-4963-9563-4943a584398c-kube-api-access-7vmj7\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.392249 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"5ef242c1-0127-4963-9563-4943a584398c\") " pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.591250 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.724670 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.862996 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2172dfb5-c0a6-4525-a326-dd9736eb5233-rabbitmq-confd\") pod \"2172dfb5-c0a6-4525-a326-dd9736eb5233\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.863322 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rprv\" (UniqueName: \"kubernetes.io/projected/2172dfb5-c0a6-4525-a326-dd9736eb5233-kube-api-access-4rprv\") pod \"2172dfb5-c0a6-4525-a326-dd9736eb5233\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.863364 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2172dfb5-c0a6-4525-a326-dd9736eb5233-rabbitmq-plugins\") pod \"2172dfb5-c0a6-4525-a326-dd9736eb5233\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.863411 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2172dfb5-c0a6-4525-a326-dd9736eb5233-rabbitmq-erlang-cookie\") pod \"2172dfb5-c0a6-4525-a326-dd9736eb5233\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.863428 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2172dfb5-c0a6-4525-a326-dd9736eb5233-erlang-cookie-secret\") pod \"2172dfb5-c0a6-4525-a326-dd9736eb5233\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.863454 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2172dfb5-c0a6-4525-a326-dd9736eb5233-rabbitmq-tls\") pod \"2172dfb5-c0a6-4525-a326-dd9736eb5233\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.863522 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"2172dfb5-c0a6-4525-a326-dd9736eb5233\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.863545 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2172dfb5-c0a6-4525-a326-dd9736eb5233-server-conf\") pod \"2172dfb5-c0a6-4525-a326-dd9736eb5233\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.863566 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2172dfb5-c0a6-4525-a326-dd9736eb5233-plugins-conf\") pod \"2172dfb5-c0a6-4525-a326-dd9736eb5233\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.863623 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2172dfb5-c0a6-4525-a326-dd9736eb5233-config-data\") pod \"2172dfb5-c0a6-4525-a326-dd9736eb5233\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.863671 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2172dfb5-c0a6-4525-a326-dd9736eb5233-pod-info\") pod \"2172dfb5-c0a6-4525-a326-dd9736eb5233\" (UID: \"2172dfb5-c0a6-4525-a326-dd9736eb5233\") " Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.863772 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2172dfb5-c0a6-4525-a326-dd9736eb5233-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "2172dfb5-c0a6-4525-a326-dd9736eb5233" (UID: "2172dfb5-c0a6-4525-a326-dd9736eb5233"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.863996 4813 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2172dfb5-c0a6-4525-a326-dd9736eb5233-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.870530 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2172dfb5-c0a6-4525-a326-dd9736eb5233-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "2172dfb5-c0a6-4525-a326-dd9736eb5233" (UID: "2172dfb5-c0a6-4525-a326-dd9736eb5233"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.878646 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2172dfb5-c0a6-4525-a326-dd9736eb5233-kube-api-access-4rprv" (OuterVolumeSpecName: "kube-api-access-4rprv") pod "2172dfb5-c0a6-4525-a326-dd9736eb5233" (UID: "2172dfb5-c0a6-4525-a326-dd9736eb5233"). InnerVolumeSpecName "kube-api-access-4rprv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.878969 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "2172dfb5-c0a6-4525-a326-dd9736eb5233" (UID: "2172dfb5-c0a6-4525-a326-dd9736eb5233"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.880033 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2172dfb5-c0a6-4525-a326-dd9736eb5233-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "2172dfb5-c0a6-4525-a326-dd9736eb5233" (UID: "2172dfb5-c0a6-4525-a326-dd9736eb5233"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.880780 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2172dfb5-c0a6-4525-a326-dd9736eb5233-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "2172dfb5-c0a6-4525-a326-dd9736eb5233" (UID: "2172dfb5-c0a6-4525-a326-dd9736eb5233"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.882124 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2172dfb5-c0a6-4525-a326-dd9736eb5233-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "2172dfb5-c0a6-4525-a326-dd9736eb5233" (UID: "2172dfb5-c0a6-4525-a326-dd9736eb5233"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.884388 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/2172dfb5-c0a6-4525-a326-dd9736eb5233-pod-info" (OuterVolumeSpecName: "pod-info") pod "2172dfb5-c0a6-4525-a326-dd9736eb5233" (UID: "2172dfb5-c0a6-4525-a326-dd9736eb5233"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.885838 4813 generic.go:334] "Generic (PLEG): container finished" podID="2172dfb5-c0a6-4525-a326-dd9736eb5233" containerID="43a1c843efdf2356e06677d802b023e646e54b4c679dfa4d791704c633a487bd" exitCode=0 Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.885886 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2172dfb5-c0a6-4525-a326-dd9736eb5233","Type":"ContainerDied","Data":"43a1c843efdf2356e06677d802b023e646e54b4c679dfa4d791704c633a487bd"} Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.885912 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2172dfb5-c0a6-4525-a326-dd9736eb5233","Type":"ContainerDied","Data":"46805f68b4ee42a603c02323712604516589877604fd0908587d6bc9a4a4616f"} Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.885929 4813 scope.go:117] "RemoveContainer" containerID="43a1c843efdf2356e06677d802b023e646e54b4c679dfa4d791704c633a487bd" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.885998 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.915124 4813 scope.go:117] "RemoveContainer" containerID="2a1916b4fe499237d82c88926954303ec3c62edb0721baeedf52bceebb68441b" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.936455 4813 scope.go:117] "RemoveContainer" containerID="43a1c843efdf2356e06677d802b023e646e54b4c679dfa4d791704c633a487bd" Dec 06 16:06:57 crc kubenswrapper[4813]: E1206 16:06:57.936985 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43a1c843efdf2356e06677d802b023e646e54b4c679dfa4d791704c633a487bd\": container with ID starting with 43a1c843efdf2356e06677d802b023e646e54b4c679dfa4d791704c633a487bd not found: ID does not exist" containerID="43a1c843efdf2356e06677d802b023e646e54b4c679dfa4d791704c633a487bd" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.937028 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43a1c843efdf2356e06677d802b023e646e54b4c679dfa4d791704c633a487bd"} err="failed to get container status \"43a1c843efdf2356e06677d802b023e646e54b4c679dfa4d791704c633a487bd\": rpc error: code = NotFound desc = could not find container \"43a1c843efdf2356e06677d802b023e646e54b4c679dfa4d791704c633a487bd\": container with ID starting with 43a1c843efdf2356e06677d802b023e646e54b4c679dfa4d791704c633a487bd not found: ID does not exist" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.937051 4813 scope.go:117] "RemoveContainer" containerID="2a1916b4fe499237d82c88926954303ec3c62edb0721baeedf52bceebb68441b" Dec 06 16:06:57 crc kubenswrapper[4813]: E1206 16:06:57.937405 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a1916b4fe499237d82c88926954303ec3c62edb0721baeedf52bceebb68441b\": container with ID starting with 2a1916b4fe499237d82c88926954303ec3c62edb0721baeedf52bceebb68441b not found: ID does not exist" containerID="2a1916b4fe499237d82c88926954303ec3c62edb0721baeedf52bceebb68441b" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.937438 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a1916b4fe499237d82c88926954303ec3c62edb0721baeedf52bceebb68441b"} err="failed to get container status \"2a1916b4fe499237d82c88926954303ec3c62edb0721baeedf52bceebb68441b\": rpc error: code = NotFound desc = could not find container \"2a1916b4fe499237d82c88926954303ec3c62edb0721baeedf52bceebb68441b\": container with ID starting with 2a1916b4fe499237d82c88926954303ec3c62edb0721baeedf52bceebb68441b not found: ID does not exist" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.938346 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2172dfb5-c0a6-4525-a326-dd9736eb5233-config-data" (OuterVolumeSpecName: "config-data") pod "2172dfb5-c0a6-4525-a326-dd9736eb5233" (UID: "2172dfb5-c0a6-4525-a326-dd9736eb5233"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.966088 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.966118 4813 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2172dfb5-c0a6-4525-a326-dd9736eb5233-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.966128 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2172dfb5-c0a6-4525-a326-dd9736eb5233-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.966137 4813 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2172dfb5-c0a6-4525-a326-dd9736eb5233-pod-info\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.966146 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rprv\" (UniqueName: \"kubernetes.io/projected/2172dfb5-c0a6-4525-a326-dd9736eb5233-kube-api-access-4rprv\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.966155 4813 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2172dfb5-c0a6-4525-a326-dd9736eb5233-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.966163 4813 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2172dfb5-c0a6-4525-a326-dd9736eb5233-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.966172 4813 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2172dfb5-c0a6-4525-a326-dd9736eb5233-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.986023 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2172dfb5-c0a6-4525-a326-dd9736eb5233-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "2172dfb5-c0a6-4525-a326-dd9736eb5233" (UID: "2172dfb5-c0a6-4525-a326-dd9736eb5233"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.992831 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2172dfb5-c0a6-4525-a326-dd9736eb5233-server-conf" (OuterVolumeSpecName: "server-conf") pod "2172dfb5-c0a6-4525-a326-dd9736eb5233" (UID: "2172dfb5-c0a6-4525-a326-dd9736eb5233"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:06:57 crc kubenswrapper[4813]: I1206 16:06:57.994086 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.069396 4813 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2172dfb5-c0a6-4525-a326-dd9736eb5233-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.069428 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.069448 4813 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2172dfb5-c0a6-4525-a326-dd9736eb5233-server-conf\") on node \"crc\" DevicePath \"\"" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.071497 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.215128 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.223539 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.249222 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 16:06:58 crc kubenswrapper[4813]: E1206 16:06:58.249612 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2172dfb5-c0a6-4525-a326-dd9736eb5233" containerName="rabbitmq" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.249629 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="2172dfb5-c0a6-4525-a326-dd9736eb5233" containerName="rabbitmq" Dec 06 16:06:58 crc kubenswrapper[4813]: E1206 16:06:58.249671 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2172dfb5-c0a6-4525-a326-dd9736eb5233" containerName="setup-container" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.249678 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="2172dfb5-c0a6-4525-a326-dd9736eb5233" containerName="setup-container" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.249851 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="2172dfb5-c0a6-4525-a326-dd9736eb5233" containerName="rabbitmq" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.250778 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.254468 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.255021 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-69ctp" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.255061 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.255915 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.256973 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.257390 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.257654 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.270697 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.374016 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/342df504-7a38-4c50-bb3d-8c7b28521e64-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.374271 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.374300 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/342df504-7a38-4c50-bb3d-8c7b28521e64-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.374327 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/342df504-7a38-4c50-bb3d-8c7b28521e64-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.374352 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8g2j\" (UniqueName: \"kubernetes.io/projected/342df504-7a38-4c50-bb3d-8c7b28521e64-kube-api-access-t8g2j\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.374600 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/342df504-7a38-4c50-bb3d-8c7b28521e64-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.374744 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/342df504-7a38-4c50-bb3d-8c7b28521e64-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.374807 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/342df504-7a38-4c50-bb3d-8c7b28521e64-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.374879 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/342df504-7a38-4c50-bb3d-8c7b28521e64-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.374919 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/342df504-7a38-4c50-bb3d-8c7b28521e64-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.374955 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/342df504-7a38-4c50-bb3d-8c7b28521e64-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.476724 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/342df504-7a38-4c50-bb3d-8c7b28521e64-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.476786 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/342df504-7a38-4c50-bb3d-8c7b28521e64-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.476812 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/342df504-7a38-4c50-bb3d-8c7b28521e64-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.476830 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/342df504-7a38-4c50-bb3d-8c7b28521e64-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.476850 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/342df504-7a38-4c50-bb3d-8c7b28521e64-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.476884 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/342df504-7a38-4c50-bb3d-8c7b28521e64-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.476923 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/342df504-7a38-4c50-bb3d-8c7b28521e64-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.476939 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.476960 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/342df504-7a38-4c50-bb3d-8c7b28521e64-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.476982 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/342df504-7a38-4c50-bb3d-8c7b28521e64-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.477006 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8g2j\" (UniqueName: \"kubernetes.io/projected/342df504-7a38-4c50-bb3d-8c7b28521e64-kube-api-access-t8g2j\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.477498 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.477889 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/342df504-7a38-4c50-bb3d-8c7b28521e64-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.478045 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/342df504-7a38-4c50-bb3d-8c7b28521e64-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.479277 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/342df504-7a38-4c50-bb3d-8c7b28521e64-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.479603 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/342df504-7a38-4c50-bb3d-8c7b28521e64-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.482038 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/342df504-7a38-4c50-bb3d-8c7b28521e64-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.483989 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/342df504-7a38-4c50-bb3d-8c7b28521e64-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.484502 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/342df504-7a38-4c50-bb3d-8c7b28521e64-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.488968 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/342df504-7a38-4c50-bb3d-8c7b28521e64-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.489115 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/342df504-7a38-4c50-bb3d-8c7b28521e64-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.493113 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8g2j\" (UniqueName: \"kubernetes.io/projected/342df504-7a38-4c50-bb3d-8c7b28521e64-kube-api-access-t8g2j\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.501994 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2172dfb5-c0a6-4525-a326-dd9736eb5233" path="/var/lib/kubelet/pods/2172dfb5-c0a6-4525-a326-dd9736eb5233/volumes" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.503516 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f03c3251-5e1c-4cb2-aa3f-ebbef28acb19" path="/var/lib/kubelet/pods/f03c3251-5e1c-4cb2-aa3f-ebbef28acb19/volumes" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.513223 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"342df504-7a38-4c50-bb3d-8c7b28521e64\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.630810 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:06:58 crc kubenswrapper[4813]: I1206 16:06:58.896334 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5ef242c1-0127-4963-9563-4943a584398c","Type":"ContainerStarted","Data":"adb9c0b2755fddd3ba6f100ed6afb5ab421f17e6d16f5afcefa2a5a478c97816"} Dec 06 16:06:59 crc kubenswrapper[4813]: I1206 16:06:59.118623 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 16:06:59 crc kubenswrapper[4813]: W1206 16:06:59.132875 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod342df504_7a38_4c50_bb3d_8c7b28521e64.slice/crio-9bf7646381999115d3cf72f4a2de1acfe4b727a5e43d12ed401221bad688a80a WatchSource:0}: Error finding container 9bf7646381999115d3cf72f4a2de1acfe4b727a5e43d12ed401221bad688a80a: Status 404 returned error can't find the container with id 9bf7646381999115d3cf72f4a2de1acfe4b727a5e43d12ed401221bad688a80a Dec 06 16:06:59 crc kubenswrapper[4813]: I1206 16:06:59.780540 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d558885bc-qlbfw"] Dec 06 16:06:59 crc kubenswrapper[4813]: I1206 16:06:59.783573 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-qlbfw" Dec 06 16:06:59 crc kubenswrapper[4813]: I1206 16:06:59.786460 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 06 16:06:59 crc kubenswrapper[4813]: I1206 16:06:59.789249 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-qlbfw"] Dec 06 16:06:59 crc kubenswrapper[4813]: I1206 16:06:59.902591 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-config\") pod \"dnsmasq-dns-d558885bc-qlbfw\" (UID: \"14088ff7-a199-46ee-8bb7-61fd08539075\") " pod="openstack/dnsmasq-dns-d558885bc-qlbfw" Dec 06 16:06:59 crc kubenswrapper[4813]: I1206 16:06:59.902690 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-qlbfw\" (UID: \"14088ff7-a199-46ee-8bb7-61fd08539075\") " pod="openstack/dnsmasq-dns-d558885bc-qlbfw" Dec 06 16:06:59 crc kubenswrapper[4813]: I1206 16:06:59.902730 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-qlbfw\" (UID: \"14088ff7-a199-46ee-8bb7-61fd08539075\") " pod="openstack/dnsmasq-dns-d558885bc-qlbfw" Dec 06 16:06:59 crc kubenswrapper[4813]: I1206 16:06:59.902751 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-dns-svc\") pod \"dnsmasq-dns-d558885bc-qlbfw\" (UID: \"14088ff7-a199-46ee-8bb7-61fd08539075\") " pod="openstack/dnsmasq-dns-d558885bc-qlbfw" Dec 06 16:06:59 crc kubenswrapper[4813]: I1206 16:06:59.902944 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-qlbfw\" (UID: \"14088ff7-a199-46ee-8bb7-61fd08539075\") " pod="openstack/dnsmasq-dns-d558885bc-qlbfw" Dec 06 16:06:59 crc kubenswrapper[4813]: I1206 16:06:59.905890 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdhpm\" (UniqueName: \"kubernetes.io/projected/14088ff7-a199-46ee-8bb7-61fd08539075-kube-api-access-zdhpm\") pod \"dnsmasq-dns-d558885bc-qlbfw\" (UID: \"14088ff7-a199-46ee-8bb7-61fd08539075\") " pod="openstack/dnsmasq-dns-d558885bc-qlbfw" Dec 06 16:06:59 crc kubenswrapper[4813]: I1206 16:06:59.905964 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-qlbfw\" (UID: \"14088ff7-a199-46ee-8bb7-61fd08539075\") " pod="openstack/dnsmasq-dns-d558885bc-qlbfw" Dec 06 16:06:59 crc kubenswrapper[4813]: I1206 16:06:59.907520 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"342df504-7a38-4c50-bb3d-8c7b28521e64","Type":"ContainerStarted","Data":"9bf7646381999115d3cf72f4a2de1acfe4b727a5e43d12ed401221bad688a80a"} Dec 06 16:06:59 crc kubenswrapper[4813]: I1206 16:06:59.909234 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5ef242c1-0127-4963-9563-4943a584398c","Type":"ContainerStarted","Data":"8c1a9416fb0187497d5d1ff7e0cd529f35d5d2af505ea0ca253248d032f305d9"} Dec 06 16:07:00 crc kubenswrapper[4813]: I1206 16:07:00.008338 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-qlbfw\" (UID: \"14088ff7-a199-46ee-8bb7-61fd08539075\") " pod="openstack/dnsmasq-dns-d558885bc-qlbfw" Dec 06 16:07:00 crc kubenswrapper[4813]: I1206 16:07:00.008530 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdhpm\" (UniqueName: \"kubernetes.io/projected/14088ff7-a199-46ee-8bb7-61fd08539075-kube-api-access-zdhpm\") pod \"dnsmasq-dns-d558885bc-qlbfw\" (UID: \"14088ff7-a199-46ee-8bb7-61fd08539075\") " pod="openstack/dnsmasq-dns-d558885bc-qlbfw" Dec 06 16:07:00 crc kubenswrapper[4813]: I1206 16:07:00.008571 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-qlbfw\" (UID: \"14088ff7-a199-46ee-8bb7-61fd08539075\") " pod="openstack/dnsmasq-dns-d558885bc-qlbfw" Dec 06 16:07:00 crc kubenswrapper[4813]: I1206 16:07:00.008667 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-config\") pod \"dnsmasq-dns-d558885bc-qlbfw\" (UID: \"14088ff7-a199-46ee-8bb7-61fd08539075\") " pod="openstack/dnsmasq-dns-d558885bc-qlbfw" Dec 06 16:07:00 crc kubenswrapper[4813]: I1206 16:07:00.008717 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-qlbfw\" (UID: \"14088ff7-a199-46ee-8bb7-61fd08539075\") " pod="openstack/dnsmasq-dns-d558885bc-qlbfw" Dec 06 16:07:00 crc kubenswrapper[4813]: I1206 16:07:00.008762 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-qlbfw\" (UID: \"14088ff7-a199-46ee-8bb7-61fd08539075\") " pod="openstack/dnsmasq-dns-d558885bc-qlbfw" Dec 06 16:07:00 crc kubenswrapper[4813]: I1206 16:07:00.008794 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-dns-svc\") pod \"dnsmasq-dns-d558885bc-qlbfw\" (UID: \"14088ff7-a199-46ee-8bb7-61fd08539075\") " pod="openstack/dnsmasq-dns-d558885bc-qlbfw" Dec 06 16:07:00 crc kubenswrapper[4813]: I1206 16:07:00.010648 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-qlbfw\" (UID: \"14088ff7-a199-46ee-8bb7-61fd08539075\") " pod="openstack/dnsmasq-dns-d558885bc-qlbfw" Dec 06 16:07:00 crc kubenswrapper[4813]: I1206 16:07:00.010838 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-qlbfw\" (UID: \"14088ff7-a199-46ee-8bb7-61fd08539075\") " pod="openstack/dnsmasq-dns-d558885bc-qlbfw" Dec 06 16:07:00 crc kubenswrapper[4813]: I1206 16:07:00.010858 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-config\") pod \"dnsmasq-dns-d558885bc-qlbfw\" (UID: \"14088ff7-a199-46ee-8bb7-61fd08539075\") " pod="openstack/dnsmasq-dns-d558885bc-qlbfw" Dec 06 16:07:00 crc kubenswrapper[4813]: I1206 16:07:00.011378 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-qlbfw\" (UID: \"14088ff7-a199-46ee-8bb7-61fd08539075\") " pod="openstack/dnsmasq-dns-d558885bc-qlbfw" Dec 06 16:07:00 crc kubenswrapper[4813]: I1206 16:07:00.011388 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-qlbfw\" (UID: \"14088ff7-a199-46ee-8bb7-61fd08539075\") " pod="openstack/dnsmasq-dns-d558885bc-qlbfw" Dec 06 16:07:00 crc kubenswrapper[4813]: I1206 16:07:00.011569 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-dns-svc\") pod \"dnsmasq-dns-d558885bc-qlbfw\" (UID: \"14088ff7-a199-46ee-8bb7-61fd08539075\") " pod="openstack/dnsmasq-dns-d558885bc-qlbfw" Dec 06 16:07:00 crc kubenswrapper[4813]: I1206 16:07:00.041956 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdhpm\" (UniqueName: \"kubernetes.io/projected/14088ff7-a199-46ee-8bb7-61fd08539075-kube-api-access-zdhpm\") pod \"dnsmasq-dns-d558885bc-qlbfw\" (UID: \"14088ff7-a199-46ee-8bb7-61fd08539075\") " pod="openstack/dnsmasq-dns-d558885bc-qlbfw" Dec 06 16:07:00 crc kubenswrapper[4813]: I1206 16:07:00.137642 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-qlbfw" Dec 06 16:07:00 crc kubenswrapper[4813]: I1206 16:07:00.590477 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-qlbfw"] Dec 06 16:07:00 crc kubenswrapper[4813]: W1206 16:07:00.607746 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod14088ff7_a199_46ee_8bb7_61fd08539075.slice/crio-2d578c0ad378187d85a274b57f154a00e4c2cdfae27f04318e4df2850176e604 WatchSource:0}: Error finding container 2d578c0ad378187d85a274b57f154a00e4c2cdfae27f04318e4df2850176e604: Status 404 returned error can't find the container with id 2d578c0ad378187d85a274b57f154a00e4c2cdfae27f04318e4df2850176e604 Dec 06 16:07:00 crc kubenswrapper[4813]: I1206 16:07:00.918345 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-qlbfw" event={"ID":"14088ff7-a199-46ee-8bb7-61fd08539075","Type":"ContainerStarted","Data":"2d578c0ad378187d85a274b57f154a00e4c2cdfae27f04318e4df2850176e604"} Dec 06 16:07:01 crc kubenswrapper[4813]: I1206 16:07:01.951019 4813 generic.go:334] "Generic (PLEG): container finished" podID="14088ff7-a199-46ee-8bb7-61fd08539075" containerID="cff0d75209ccaf97a4bbfe292a334b8a81c501f9234c7c9c200068afee1aa67f" exitCode=0 Dec 06 16:07:01 crc kubenswrapper[4813]: I1206 16:07:01.951220 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-qlbfw" event={"ID":"14088ff7-a199-46ee-8bb7-61fd08539075","Type":"ContainerDied","Data":"cff0d75209ccaf97a4bbfe292a334b8a81c501f9234c7c9c200068afee1aa67f"} Dec 06 16:07:01 crc kubenswrapper[4813]: I1206 16:07:01.955188 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"342df504-7a38-4c50-bb3d-8c7b28521e64","Type":"ContainerStarted","Data":"38a3f4293c730c5712b5ec5130a65f4f6b714b7d1e2ec48fc5687a010449734a"} Dec 06 16:07:02 crc kubenswrapper[4813]: I1206 16:07:02.971775 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-qlbfw" event={"ID":"14088ff7-a199-46ee-8bb7-61fd08539075","Type":"ContainerStarted","Data":"4da197882caa74ad27c195e8c935cbe2bd78b309d605cdcd4cdf3f3dc1b8325b"} Dec 06 16:07:02 crc kubenswrapper[4813]: I1206 16:07:02.972591 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d558885bc-qlbfw" Dec 06 16:07:03 crc kubenswrapper[4813]: I1206 16:07:03.012190 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d558885bc-qlbfw" podStartSLOduration=4.012165225 podStartE2EDuration="4.012165225s" podCreationTimestamp="2025-12-06 16:06:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:07:03.002988103 +0000 UTC m=+1262.893867729" watchObservedRunningTime="2025-12-06 16:07:03.012165225 +0000 UTC m=+1262.903044841" Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.139444 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d558885bc-qlbfw" Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.223567 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-8rcjh"] Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.223801 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" podUID="df5ab4f0-9f97-47a3-89d0-ef1680c215e7" containerName="dnsmasq-dns" containerID="cri-o://ed00f85126577ceec9342dac944664aacece3dd1fce9976f8bfaf8815e5cd603" gracePeriod=10 Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.518406 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b6dc74c5-4w85d"] Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.529818 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b6dc74c5-4w85d"] Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.529927 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b6dc74c5-4w85d" Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.650169 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5399f130-d1b7-48ba-a32b-6da83416719e-ovsdbserver-nb\") pod \"dnsmasq-dns-6b6dc74c5-4w85d\" (UID: \"5399f130-d1b7-48ba-a32b-6da83416719e\") " pod="openstack/dnsmasq-dns-6b6dc74c5-4w85d" Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.650240 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5399f130-d1b7-48ba-a32b-6da83416719e-dns-swift-storage-0\") pod \"dnsmasq-dns-6b6dc74c5-4w85d\" (UID: \"5399f130-d1b7-48ba-a32b-6da83416719e\") " pod="openstack/dnsmasq-dns-6b6dc74c5-4w85d" Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.650289 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjmz5\" (UniqueName: \"kubernetes.io/projected/5399f130-d1b7-48ba-a32b-6da83416719e-kube-api-access-vjmz5\") pod \"dnsmasq-dns-6b6dc74c5-4w85d\" (UID: \"5399f130-d1b7-48ba-a32b-6da83416719e\") " pod="openstack/dnsmasq-dns-6b6dc74c5-4w85d" Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.650532 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5399f130-d1b7-48ba-a32b-6da83416719e-config\") pod \"dnsmasq-dns-6b6dc74c5-4w85d\" (UID: \"5399f130-d1b7-48ba-a32b-6da83416719e\") " pod="openstack/dnsmasq-dns-6b6dc74c5-4w85d" Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.650607 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5399f130-d1b7-48ba-a32b-6da83416719e-openstack-edpm-ipam\") pod \"dnsmasq-dns-6b6dc74c5-4w85d\" (UID: \"5399f130-d1b7-48ba-a32b-6da83416719e\") " pod="openstack/dnsmasq-dns-6b6dc74c5-4w85d" Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.650757 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5399f130-d1b7-48ba-a32b-6da83416719e-ovsdbserver-sb\") pod \"dnsmasq-dns-6b6dc74c5-4w85d\" (UID: \"5399f130-d1b7-48ba-a32b-6da83416719e\") " pod="openstack/dnsmasq-dns-6b6dc74c5-4w85d" Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.650825 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5399f130-d1b7-48ba-a32b-6da83416719e-dns-svc\") pod \"dnsmasq-dns-6b6dc74c5-4w85d\" (UID: \"5399f130-d1b7-48ba-a32b-6da83416719e\") " pod="openstack/dnsmasq-dns-6b6dc74c5-4w85d" Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.752421 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5399f130-d1b7-48ba-a32b-6da83416719e-ovsdbserver-nb\") pod \"dnsmasq-dns-6b6dc74c5-4w85d\" (UID: \"5399f130-d1b7-48ba-a32b-6da83416719e\") " pod="openstack/dnsmasq-dns-6b6dc74c5-4w85d" Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.752484 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5399f130-d1b7-48ba-a32b-6da83416719e-dns-swift-storage-0\") pod \"dnsmasq-dns-6b6dc74c5-4w85d\" (UID: \"5399f130-d1b7-48ba-a32b-6da83416719e\") " pod="openstack/dnsmasq-dns-6b6dc74c5-4w85d" Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.752500 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjmz5\" (UniqueName: \"kubernetes.io/projected/5399f130-d1b7-48ba-a32b-6da83416719e-kube-api-access-vjmz5\") pod \"dnsmasq-dns-6b6dc74c5-4w85d\" (UID: \"5399f130-d1b7-48ba-a32b-6da83416719e\") " pod="openstack/dnsmasq-dns-6b6dc74c5-4w85d" Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.752556 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5399f130-d1b7-48ba-a32b-6da83416719e-config\") pod \"dnsmasq-dns-6b6dc74c5-4w85d\" (UID: \"5399f130-d1b7-48ba-a32b-6da83416719e\") " pod="openstack/dnsmasq-dns-6b6dc74c5-4w85d" Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.752582 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5399f130-d1b7-48ba-a32b-6da83416719e-openstack-edpm-ipam\") pod \"dnsmasq-dns-6b6dc74c5-4w85d\" (UID: \"5399f130-d1b7-48ba-a32b-6da83416719e\") " pod="openstack/dnsmasq-dns-6b6dc74c5-4w85d" Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.752630 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5399f130-d1b7-48ba-a32b-6da83416719e-ovsdbserver-sb\") pod \"dnsmasq-dns-6b6dc74c5-4w85d\" (UID: \"5399f130-d1b7-48ba-a32b-6da83416719e\") " pod="openstack/dnsmasq-dns-6b6dc74c5-4w85d" Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.752660 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5399f130-d1b7-48ba-a32b-6da83416719e-dns-svc\") pod \"dnsmasq-dns-6b6dc74c5-4w85d\" (UID: \"5399f130-d1b7-48ba-a32b-6da83416719e\") " pod="openstack/dnsmasq-dns-6b6dc74c5-4w85d" Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.754236 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5399f130-d1b7-48ba-a32b-6da83416719e-dns-swift-storage-0\") pod \"dnsmasq-dns-6b6dc74c5-4w85d\" (UID: \"5399f130-d1b7-48ba-a32b-6da83416719e\") " pod="openstack/dnsmasq-dns-6b6dc74c5-4w85d" Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.754246 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5399f130-d1b7-48ba-a32b-6da83416719e-config\") pod \"dnsmasq-dns-6b6dc74c5-4w85d\" (UID: \"5399f130-d1b7-48ba-a32b-6da83416719e\") " pod="openstack/dnsmasq-dns-6b6dc74c5-4w85d" Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.754914 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5399f130-d1b7-48ba-a32b-6da83416719e-ovsdbserver-sb\") pod \"dnsmasq-dns-6b6dc74c5-4w85d\" (UID: \"5399f130-d1b7-48ba-a32b-6da83416719e\") " pod="openstack/dnsmasq-dns-6b6dc74c5-4w85d" Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.754917 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5399f130-d1b7-48ba-a32b-6da83416719e-ovsdbserver-nb\") pod \"dnsmasq-dns-6b6dc74c5-4w85d\" (UID: \"5399f130-d1b7-48ba-a32b-6da83416719e\") " pod="openstack/dnsmasq-dns-6b6dc74c5-4w85d" Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.755471 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5399f130-d1b7-48ba-a32b-6da83416719e-dns-svc\") pod \"dnsmasq-dns-6b6dc74c5-4w85d\" (UID: \"5399f130-d1b7-48ba-a32b-6da83416719e\") " pod="openstack/dnsmasq-dns-6b6dc74c5-4w85d" Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.755765 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5399f130-d1b7-48ba-a32b-6da83416719e-openstack-edpm-ipam\") pod \"dnsmasq-dns-6b6dc74c5-4w85d\" (UID: \"5399f130-d1b7-48ba-a32b-6da83416719e\") " pod="openstack/dnsmasq-dns-6b6dc74c5-4w85d" Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.770876 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjmz5\" (UniqueName: \"kubernetes.io/projected/5399f130-d1b7-48ba-a32b-6da83416719e-kube-api-access-vjmz5\") pod \"dnsmasq-dns-6b6dc74c5-4w85d\" (UID: \"5399f130-d1b7-48ba-a32b-6da83416719e\") " pod="openstack/dnsmasq-dns-6b6dc74c5-4w85d" Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.829435 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.880143 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b6dc74c5-4w85d" Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.956101 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-ovsdbserver-nb\") pod \"df5ab4f0-9f97-47a3-89d0-ef1680c215e7\" (UID: \"df5ab4f0-9f97-47a3-89d0-ef1680c215e7\") " Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.956156 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-config\") pod \"df5ab4f0-9f97-47a3-89d0-ef1680c215e7\" (UID: \"df5ab4f0-9f97-47a3-89d0-ef1680c215e7\") " Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.956301 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-ovsdbserver-sb\") pod \"df5ab4f0-9f97-47a3-89d0-ef1680c215e7\" (UID: \"df5ab4f0-9f97-47a3-89d0-ef1680c215e7\") " Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.956325 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-dns-svc\") pod \"df5ab4f0-9f97-47a3-89d0-ef1680c215e7\" (UID: \"df5ab4f0-9f97-47a3-89d0-ef1680c215e7\") " Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.956355 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-dns-swift-storage-0\") pod \"df5ab4f0-9f97-47a3-89d0-ef1680c215e7\" (UID: \"df5ab4f0-9f97-47a3-89d0-ef1680c215e7\") " Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.956411 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ws8wh\" (UniqueName: \"kubernetes.io/projected/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-kube-api-access-ws8wh\") pod \"df5ab4f0-9f97-47a3-89d0-ef1680c215e7\" (UID: \"df5ab4f0-9f97-47a3-89d0-ef1680c215e7\") " Dec 06 16:07:10 crc kubenswrapper[4813]: I1206 16:07:10.974075 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-kube-api-access-ws8wh" (OuterVolumeSpecName: "kube-api-access-ws8wh") pod "df5ab4f0-9f97-47a3-89d0-ef1680c215e7" (UID: "df5ab4f0-9f97-47a3-89d0-ef1680c215e7"). InnerVolumeSpecName "kube-api-access-ws8wh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:07:11 crc kubenswrapper[4813]: I1206 16:07:11.038466 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "df5ab4f0-9f97-47a3-89d0-ef1680c215e7" (UID: "df5ab4f0-9f97-47a3-89d0-ef1680c215e7"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:07:11 crc kubenswrapper[4813]: I1206 16:07:11.046973 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-config" (OuterVolumeSpecName: "config") pod "df5ab4f0-9f97-47a3-89d0-ef1680c215e7" (UID: "df5ab4f0-9f97-47a3-89d0-ef1680c215e7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:07:11 crc kubenswrapper[4813]: I1206 16:07:11.060760 4813 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 16:07:11 crc kubenswrapper[4813]: I1206 16:07:11.060780 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ws8wh\" (UniqueName: \"kubernetes.io/projected/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-kube-api-access-ws8wh\") on node \"crc\" DevicePath \"\"" Dec 06 16:07:11 crc kubenswrapper[4813]: I1206 16:07:11.060791 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-config\") on node \"crc\" DevicePath \"\"" Dec 06 16:07:11 crc kubenswrapper[4813]: I1206 16:07:11.061745 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "df5ab4f0-9f97-47a3-89d0-ef1680c215e7" (UID: "df5ab4f0-9f97-47a3-89d0-ef1680c215e7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:07:11 crc kubenswrapper[4813]: I1206 16:07:11.063858 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "df5ab4f0-9f97-47a3-89d0-ef1680c215e7" (UID: "df5ab4f0-9f97-47a3-89d0-ef1680c215e7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:07:11 crc kubenswrapper[4813]: I1206 16:07:11.068320 4813 generic.go:334] "Generic (PLEG): container finished" podID="df5ab4f0-9f97-47a3-89d0-ef1680c215e7" containerID="ed00f85126577ceec9342dac944664aacece3dd1fce9976f8bfaf8815e5cd603" exitCode=0 Dec 06 16:07:11 crc kubenswrapper[4813]: I1206 16:07:11.068350 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" event={"ID":"df5ab4f0-9f97-47a3-89d0-ef1680c215e7","Type":"ContainerDied","Data":"ed00f85126577ceec9342dac944664aacece3dd1fce9976f8bfaf8815e5cd603"} Dec 06 16:07:11 crc kubenswrapper[4813]: I1206 16:07:11.068375 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" event={"ID":"df5ab4f0-9f97-47a3-89d0-ef1680c215e7","Type":"ContainerDied","Data":"25a7109e7cbea43e65fda9df467a29ab5ba06c0c42e7a882c19361e16fc03d23"} Dec 06 16:07:11 crc kubenswrapper[4813]: I1206 16:07:11.068393 4813 scope.go:117] "RemoveContainer" containerID="ed00f85126577ceec9342dac944664aacece3dd1fce9976f8bfaf8815e5cd603" Dec 06 16:07:11 crc kubenswrapper[4813]: I1206 16:07:11.068514 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-8rcjh" Dec 06 16:07:11 crc kubenswrapper[4813]: I1206 16:07:11.110215 4813 scope.go:117] "RemoveContainer" containerID="f96efde1a77b21f35c209a2a0aa7f619248974fa7517d0825113a7f5383dc393" Dec 06 16:07:11 crc kubenswrapper[4813]: I1206 16:07:11.111893 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "df5ab4f0-9f97-47a3-89d0-ef1680c215e7" (UID: "df5ab4f0-9f97-47a3-89d0-ef1680c215e7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:07:11 crc kubenswrapper[4813]: I1206 16:07:11.148817 4813 scope.go:117] "RemoveContainer" containerID="ed00f85126577ceec9342dac944664aacece3dd1fce9976f8bfaf8815e5cd603" Dec 06 16:07:11 crc kubenswrapper[4813]: E1206 16:07:11.149320 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed00f85126577ceec9342dac944664aacece3dd1fce9976f8bfaf8815e5cd603\": container with ID starting with ed00f85126577ceec9342dac944664aacece3dd1fce9976f8bfaf8815e5cd603 not found: ID does not exist" containerID="ed00f85126577ceec9342dac944664aacece3dd1fce9976f8bfaf8815e5cd603" Dec 06 16:07:11 crc kubenswrapper[4813]: I1206 16:07:11.149415 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed00f85126577ceec9342dac944664aacece3dd1fce9976f8bfaf8815e5cd603"} err="failed to get container status \"ed00f85126577ceec9342dac944664aacece3dd1fce9976f8bfaf8815e5cd603\": rpc error: code = NotFound desc = could not find container \"ed00f85126577ceec9342dac944664aacece3dd1fce9976f8bfaf8815e5cd603\": container with ID starting with ed00f85126577ceec9342dac944664aacece3dd1fce9976f8bfaf8815e5cd603 not found: ID does not exist" Dec 06 16:07:11 crc kubenswrapper[4813]: I1206 16:07:11.149441 4813 scope.go:117] "RemoveContainer" containerID="f96efde1a77b21f35c209a2a0aa7f619248974fa7517d0825113a7f5383dc393" Dec 06 16:07:11 crc kubenswrapper[4813]: E1206 16:07:11.150247 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f96efde1a77b21f35c209a2a0aa7f619248974fa7517d0825113a7f5383dc393\": container with ID starting with f96efde1a77b21f35c209a2a0aa7f619248974fa7517d0825113a7f5383dc393 not found: ID does not exist" containerID="f96efde1a77b21f35c209a2a0aa7f619248974fa7517d0825113a7f5383dc393" Dec 06 16:07:11 crc kubenswrapper[4813]: I1206 16:07:11.150330 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f96efde1a77b21f35c209a2a0aa7f619248974fa7517d0825113a7f5383dc393"} err="failed to get container status \"f96efde1a77b21f35c209a2a0aa7f619248974fa7517d0825113a7f5383dc393\": rpc error: code = NotFound desc = could not find container \"f96efde1a77b21f35c209a2a0aa7f619248974fa7517d0825113a7f5383dc393\": container with ID starting with f96efde1a77b21f35c209a2a0aa7f619248974fa7517d0825113a7f5383dc393 not found: ID does not exist" Dec 06 16:07:11 crc kubenswrapper[4813]: I1206 16:07:11.162723 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 16:07:11 crc kubenswrapper[4813]: I1206 16:07:11.162757 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 16:07:11 crc kubenswrapper[4813]: I1206 16:07:11.162767 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df5ab4f0-9f97-47a3-89d0-ef1680c215e7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 16:07:11 crc kubenswrapper[4813]: I1206 16:07:11.396156 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b6dc74c5-4w85d"] Dec 06 16:07:11 crc kubenswrapper[4813]: I1206 16:07:11.554569 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-8rcjh"] Dec 06 16:07:11 crc kubenswrapper[4813]: I1206 16:07:11.568482 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-8rcjh"] Dec 06 16:07:12 crc kubenswrapper[4813]: I1206 16:07:12.080125 4813 generic.go:334] "Generic (PLEG): container finished" podID="5399f130-d1b7-48ba-a32b-6da83416719e" containerID="98a45f1e1bbde79be7b48915ea2c5f68f4f429998bbc6b8d3131db72e9523e96" exitCode=0 Dec 06 16:07:12 crc kubenswrapper[4813]: I1206 16:07:12.080436 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b6dc74c5-4w85d" event={"ID":"5399f130-d1b7-48ba-a32b-6da83416719e","Type":"ContainerDied","Data":"98a45f1e1bbde79be7b48915ea2c5f68f4f429998bbc6b8d3131db72e9523e96"} Dec 06 16:07:12 crc kubenswrapper[4813]: I1206 16:07:12.080463 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b6dc74c5-4w85d" event={"ID":"5399f130-d1b7-48ba-a32b-6da83416719e","Type":"ContainerStarted","Data":"ffe60da01e6479960c0fe6df22da3da91a98092a38c6d86a256cfa556b5701ce"} Dec 06 16:07:12 crc kubenswrapper[4813]: I1206 16:07:12.498607 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df5ab4f0-9f97-47a3-89d0-ef1680c215e7" path="/var/lib/kubelet/pods/df5ab4f0-9f97-47a3-89d0-ef1680c215e7/volumes" Dec 06 16:07:13 crc kubenswrapper[4813]: I1206 16:07:13.091542 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b6dc74c5-4w85d" event={"ID":"5399f130-d1b7-48ba-a32b-6da83416719e","Type":"ContainerStarted","Data":"286ef95f2858d08fec1443ce82b698e0f0961b808b607d6956a3907b371520cf"} Dec 06 16:07:13 crc kubenswrapper[4813]: I1206 16:07:13.091960 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b6dc74c5-4w85d" Dec 06 16:07:13 crc kubenswrapper[4813]: I1206 16:07:13.120028 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b6dc74c5-4w85d" podStartSLOduration=3.120009794 podStartE2EDuration="3.120009794s" podCreationTimestamp="2025-12-06 16:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:07:13.111653793 +0000 UTC m=+1273.002533369" watchObservedRunningTime="2025-12-06 16:07:13.120009794 +0000 UTC m=+1273.010889380" Dec 06 16:07:20 crc kubenswrapper[4813]: I1206 16:07:20.882469 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b6dc74c5-4w85d" Dec 06 16:07:20 crc kubenswrapper[4813]: I1206 16:07:20.975615 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-qlbfw"] Dec 06 16:07:20 crc kubenswrapper[4813]: I1206 16:07:20.975855 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d558885bc-qlbfw" podUID="14088ff7-a199-46ee-8bb7-61fd08539075" containerName="dnsmasq-dns" containerID="cri-o://4da197882caa74ad27c195e8c935cbe2bd78b309d605cdcd4cdf3f3dc1b8325b" gracePeriod=10 Dec 06 16:07:21 crc kubenswrapper[4813]: I1206 16:07:21.199222 4813 generic.go:334] "Generic (PLEG): container finished" podID="14088ff7-a199-46ee-8bb7-61fd08539075" containerID="4da197882caa74ad27c195e8c935cbe2bd78b309d605cdcd4cdf3f3dc1b8325b" exitCode=0 Dec 06 16:07:21 crc kubenswrapper[4813]: I1206 16:07:21.199592 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-qlbfw" event={"ID":"14088ff7-a199-46ee-8bb7-61fd08539075","Type":"ContainerDied","Data":"4da197882caa74ad27c195e8c935cbe2bd78b309d605cdcd4cdf3f3dc1b8325b"} Dec 06 16:07:21 crc kubenswrapper[4813]: I1206 16:07:21.515411 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-qlbfw" Dec 06 16:07:21 crc kubenswrapper[4813]: I1206 16:07:21.681196 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-ovsdbserver-nb\") pod \"14088ff7-a199-46ee-8bb7-61fd08539075\" (UID: \"14088ff7-a199-46ee-8bb7-61fd08539075\") " Dec 06 16:07:21 crc kubenswrapper[4813]: I1206 16:07:21.681564 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-dns-svc\") pod \"14088ff7-a199-46ee-8bb7-61fd08539075\" (UID: \"14088ff7-a199-46ee-8bb7-61fd08539075\") " Dec 06 16:07:21 crc kubenswrapper[4813]: I1206 16:07:21.681738 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdhpm\" (UniqueName: \"kubernetes.io/projected/14088ff7-a199-46ee-8bb7-61fd08539075-kube-api-access-zdhpm\") pod \"14088ff7-a199-46ee-8bb7-61fd08539075\" (UID: \"14088ff7-a199-46ee-8bb7-61fd08539075\") " Dec 06 16:07:21 crc kubenswrapper[4813]: I1206 16:07:21.681870 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-config\") pod \"14088ff7-a199-46ee-8bb7-61fd08539075\" (UID: \"14088ff7-a199-46ee-8bb7-61fd08539075\") " Dec 06 16:07:21 crc kubenswrapper[4813]: I1206 16:07:21.682292 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-ovsdbserver-sb\") pod \"14088ff7-a199-46ee-8bb7-61fd08539075\" (UID: \"14088ff7-a199-46ee-8bb7-61fd08539075\") " Dec 06 16:07:21 crc kubenswrapper[4813]: I1206 16:07:21.682472 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-dns-swift-storage-0\") pod \"14088ff7-a199-46ee-8bb7-61fd08539075\" (UID: \"14088ff7-a199-46ee-8bb7-61fd08539075\") " Dec 06 16:07:21 crc kubenswrapper[4813]: I1206 16:07:21.682835 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-openstack-edpm-ipam\") pod \"14088ff7-a199-46ee-8bb7-61fd08539075\" (UID: \"14088ff7-a199-46ee-8bb7-61fd08539075\") " Dec 06 16:07:21 crc kubenswrapper[4813]: I1206 16:07:21.702241 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14088ff7-a199-46ee-8bb7-61fd08539075-kube-api-access-zdhpm" (OuterVolumeSpecName: "kube-api-access-zdhpm") pod "14088ff7-a199-46ee-8bb7-61fd08539075" (UID: "14088ff7-a199-46ee-8bb7-61fd08539075"). InnerVolumeSpecName "kube-api-access-zdhpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:07:21 crc kubenswrapper[4813]: I1206 16:07:21.733552 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "14088ff7-a199-46ee-8bb7-61fd08539075" (UID: "14088ff7-a199-46ee-8bb7-61fd08539075"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:07:21 crc kubenswrapper[4813]: I1206 16:07:21.734780 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "14088ff7-a199-46ee-8bb7-61fd08539075" (UID: "14088ff7-a199-46ee-8bb7-61fd08539075"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:07:21 crc kubenswrapper[4813]: I1206 16:07:21.746577 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-config" (OuterVolumeSpecName: "config") pod "14088ff7-a199-46ee-8bb7-61fd08539075" (UID: "14088ff7-a199-46ee-8bb7-61fd08539075"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:07:21 crc kubenswrapper[4813]: I1206 16:07:21.753879 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "14088ff7-a199-46ee-8bb7-61fd08539075" (UID: "14088ff7-a199-46ee-8bb7-61fd08539075"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:07:21 crc kubenswrapper[4813]: I1206 16:07:21.762859 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "14088ff7-a199-46ee-8bb7-61fd08539075" (UID: "14088ff7-a199-46ee-8bb7-61fd08539075"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:07:21 crc kubenswrapper[4813]: I1206 16:07:21.764688 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "14088ff7-a199-46ee-8bb7-61fd08539075" (UID: "14088ff7-a199-46ee-8bb7-61fd08539075"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:07:21 crc kubenswrapper[4813]: I1206 16:07:21.786635 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 16:07:21 crc kubenswrapper[4813]: I1206 16:07:21.786670 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 16:07:21 crc kubenswrapper[4813]: I1206 16:07:21.786683 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdhpm\" (UniqueName: \"kubernetes.io/projected/14088ff7-a199-46ee-8bb7-61fd08539075-kube-api-access-zdhpm\") on node \"crc\" DevicePath \"\"" Dec 06 16:07:21 crc kubenswrapper[4813]: I1206 16:07:21.786693 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-config\") on node \"crc\" DevicePath \"\"" Dec 06 16:07:21 crc kubenswrapper[4813]: I1206 16:07:21.786703 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 16:07:21 crc kubenswrapper[4813]: I1206 16:07:21.786711 4813 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 16:07:21 crc kubenswrapper[4813]: I1206 16:07:21.786719 4813 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/14088ff7-a199-46ee-8bb7-61fd08539075-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 06 16:07:22 crc kubenswrapper[4813]: I1206 16:07:22.221626 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-qlbfw" event={"ID":"14088ff7-a199-46ee-8bb7-61fd08539075","Type":"ContainerDied","Data":"2d578c0ad378187d85a274b57f154a00e4c2cdfae27f04318e4df2850176e604"} Dec 06 16:07:22 crc kubenswrapper[4813]: I1206 16:07:22.222739 4813 scope.go:117] "RemoveContainer" containerID="4da197882caa74ad27c195e8c935cbe2bd78b309d605cdcd4cdf3f3dc1b8325b" Dec 06 16:07:22 crc kubenswrapper[4813]: I1206 16:07:22.221727 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-qlbfw" Dec 06 16:07:22 crc kubenswrapper[4813]: I1206 16:07:22.247696 4813 scope.go:117] "RemoveContainer" containerID="cff0d75209ccaf97a4bbfe292a334b8a81c501f9234c7c9c200068afee1aa67f" Dec 06 16:07:22 crc kubenswrapper[4813]: I1206 16:07:22.277452 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-qlbfw"] Dec 06 16:07:22 crc kubenswrapper[4813]: I1206 16:07:22.285681 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-qlbfw"] Dec 06 16:07:22 crc kubenswrapper[4813]: I1206 16:07:22.497368 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14088ff7-a199-46ee-8bb7-61fd08539075" path="/var/lib/kubelet/pods/14088ff7-a199-46ee-8bb7-61fd08539075/volumes" Dec 06 16:07:32 crc kubenswrapper[4813]: I1206 16:07:32.343225 4813 generic.go:334] "Generic (PLEG): container finished" podID="5ef242c1-0127-4963-9563-4943a584398c" containerID="8c1a9416fb0187497d5d1ff7e0cd529f35d5d2af505ea0ca253248d032f305d9" exitCode=0 Dec 06 16:07:32 crc kubenswrapper[4813]: I1206 16:07:32.343809 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5ef242c1-0127-4963-9563-4943a584398c","Type":"ContainerDied","Data":"8c1a9416fb0187497d5d1ff7e0cd529f35d5d2af505ea0ca253248d032f305d9"} Dec 06 16:07:33 crc kubenswrapper[4813]: I1206 16:07:33.354210 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5ef242c1-0127-4963-9563-4943a584398c","Type":"ContainerStarted","Data":"3f06f212a974bba7c534e2639fbbe69bbd3ffa62bf6faa52e46d6bc52db71415"} Dec 06 16:07:33 crc kubenswrapper[4813]: I1206 16:07:33.355154 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 06 16:07:33 crc kubenswrapper[4813]: I1206 16:07:33.387483 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.387464552 podStartE2EDuration="36.387464552s" podCreationTimestamp="2025-12-06 16:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:07:33.37679716 +0000 UTC m=+1293.267676736" watchObservedRunningTime="2025-12-06 16:07:33.387464552 +0000 UTC m=+1293.278344128" Dec 06 16:07:34 crc kubenswrapper[4813]: I1206 16:07:34.364879 4813 generic.go:334] "Generic (PLEG): container finished" podID="342df504-7a38-4c50-bb3d-8c7b28521e64" containerID="38a3f4293c730c5712b5ec5130a65f4f6b714b7d1e2ec48fc5687a010449734a" exitCode=0 Dec 06 16:07:34 crc kubenswrapper[4813]: I1206 16:07:34.365005 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"342df504-7a38-4c50-bb3d-8c7b28521e64","Type":"ContainerDied","Data":"38a3f4293c730c5712b5ec5130a65f4f6b714b7d1e2ec48fc5687a010449734a"} Dec 06 16:07:34 crc kubenswrapper[4813]: I1206 16:07:34.640674 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2"] Dec 06 16:07:34 crc kubenswrapper[4813]: E1206 16:07:34.641055 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df5ab4f0-9f97-47a3-89d0-ef1680c215e7" containerName="init" Dec 06 16:07:34 crc kubenswrapper[4813]: I1206 16:07:34.641071 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="df5ab4f0-9f97-47a3-89d0-ef1680c215e7" containerName="init" Dec 06 16:07:34 crc kubenswrapper[4813]: E1206 16:07:34.641096 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14088ff7-a199-46ee-8bb7-61fd08539075" containerName="init" Dec 06 16:07:34 crc kubenswrapper[4813]: I1206 16:07:34.641102 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="14088ff7-a199-46ee-8bb7-61fd08539075" containerName="init" Dec 06 16:07:34 crc kubenswrapper[4813]: E1206 16:07:34.641112 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14088ff7-a199-46ee-8bb7-61fd08539075" containerName="dnsmasq-dns" Dec 06 16:07:34 crc kubenswrapper[4813]: I1206 16:07:34.641119 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="14088ff7-a199-46ee-8bb7-61fd08539075" containerName="dnsmasq-dns" Dec 06 16:07:34 crc kubenswrapper[4813]: E1206 16:07:34.641127 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df5ab4f0-9f97-47a3-89d0-ef1680c215e7" containerName="dnsmasq-dns" Dec 06 16:07:34 crc kubenswrapper[4813]: I1206 16:07:34.641134 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="df5ab4f0-9f97-47a3-89d0-ef1680c215e7" containerName="dnsmasq-dns" Dec 06 16:07:34 crc kubenswrapper[4813]: I1206 16:07:34.641324 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="14088ff7-a199-46ee-8bb7-61fd08539075" containerName="dnsmasq-dns" Dec 06 16:07:34 crc kubenswrapper[4813]: I1206 16:07:34.641348 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="df5ab4f0-9f97-47a3-89d0-ef1680c215e7" containerName="dnsmasq-dns" Dec 06 16:07:34 crc kubenswrapper[4813]: I1206 16:07:34.641880 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2" Dec 06 16:07:34 crc kubenswrapper[4813]: I1206 16:07:34.645367 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 16:07:34 crc kubenswrapper[4813]: I1206 16:07:34.645905 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jvkhf" Dec 06 16:07:34 crc kubenswrapper[4813]: I1206 16:07:34.646176 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 16:07:34 crc kubenswrapper[4813]: I1206 16:07:34.646617 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 16:07:34 crc kubenswrapper[4813]: I1206 16:07:34.669642 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2"] Dec 06 16:07:34 crc kubenswrapper[4813]: I1206 16:07:34.762587 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef5432f4-56a6-4518-a6b1-00ac9b29293f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2\" (UID: \"ef5432f4-56a6-4518-a6b1-00ac9b29293f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2" Dec 06 16:07:34 crc kubenswrapper[4813]: I1206 16:07:34.762885 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef5432f4-56a6-4518-a6b1-00ac9b29293f-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2\" (UID: \"ef5432f4-56a6-4518-a6b1-00ac9b29293f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2" Dec 06 16:07:34 crc kubenswrapper[4813]: I1206 16:07:34.762945 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sz46d\" (UniqueName: \"kubernetes.io/projected/ef5432f4-56a6-4518-a6b1-00ac9b29293f-kube-api-access-sz46d\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2\" (UID: \"ef5432f4-56a6-4518-a6b1-00ac9b29293f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2" Dec 06 16:07:34 crc kubenswrapper[4813]: I1206 16:07:34.763005 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef5432f4-56a6-4518-a6b1-00ac9b29293f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2\" (UID: \"ef5432f4-56a6-4518-a6b1-00ac9b29293f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2" Dec 06 16:07:34 crc kubenswrapper[4813]: I1206 16:07:34.864861 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef5432f4-56a6-4518-a6b1-00ac9b29293f-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2\" (UID: \"ef5432f4-56a6-4518-a6b1-00ac9b29293f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2" Dec 06 16:07:34 crc kubenswrapper[4813]: I1206 16:07:34.864943 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sz46d\" (UniqueName: \"kubernetes.io/projected/ef5432f4-56a6-4518-a6b1-00ac9b29293f-kube-api-access-sz46d\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2\" (UID: \"ef5432f4-56a6-4518-a6b1-00ac9b29293f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2" Dec 06 16:07:34 crc kubenswrapper[4813]: I1206 16:07:34.865014 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef5432f4-56a6-4518-a6b1-00ac9b29293f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2\" (UID: \"ef5432f4-56a6-4518-a6b1-00ac9b29293f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2" Dec 06 16:07:34 crc kubenswrapper[4813]: I1206 16:07:34.865059 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef5432f4-56a6-4518-a6b1-00ac9b29293f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2\" (UID: \"ef5432f4-56a6-4518-a6b1-00ac9b29293f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2" Dec 06 16:07:34 crc kubenswrapper[4813]: I1206 16:07:34.870220 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef5432f4-56a6-4518-a6b1-00ac9b29293f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2\" (UID: \"ef5432f4-56a6-4518-a6b1-00ac9b29293f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2" Dec 06 16:07:34 crc kubenswrapper[4813]: I1206 16:07:34.872733 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef5432f4-56a6-4518-a6b1-00ac9b29293f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2\" (UID: \"ef5432f4-56a6-4518-a6b1-00ac9b29293f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2" Dec 06 16:07:34 crc kubenswrapper[4813]: I1206 16:07:34.873082 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef5432f4-56a6-4518-a6b1-00ac9b29293f-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2\" (UID: \"ef5432f4-56a6-4518-a6b1-00ac9b29293f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2" Dec 06 16:07:34 crc kubenswrapper[4813]: I1206 16:07:34.880883 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sz46d\" (UniqueName: \"kubernetes.io/projected/ef5432f4-56a6-4518-a6b1-00ac9b29293f-kube-api-access-sz46d\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2\" (UID: \"ef5432f4-56a6-4518-a6b1-00ac9b29293f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2" Dec 06 16:07:34 crc kubenswrapper[4813]: I1206 16:07:34.962781 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2" Dec 06 16:07:35 crc kubenswrapper[4813]: I1206 16:07:35.378012 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"342df504-7a38-4c50-bb3d-8c7b28521e64","Type":"ContainerStarted","Data":"4f9cfdf4283116205fcc2c9fa458a62e58e736e21bcac01a0d43c708ebc6535d"} Dec 06 16:07:35 crc kubenswrapper[4813]: I1206 16:07:35.379646 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:07:35 crc kubenswrapper[4813]: I1206 16:07:35.413960 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.41394464 podStartE2EDuration="37.41394464s" podCreationTimestamp="2025-12-06 16:06:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:07:35.40829274 +0000 UTC m=+1295.299172326" watchObservedRunningTime="2025-12-06 16:07:35.41394464 +0000 UTC m=+1295.304824226" Dec 06 16:07:35 crc kubenswrapper[4813]: I1206 16:07:35.671304 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2"] Dec 06 16:07:36 crc kubenswrapper[4813]: I1206 16:07:36.405569 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2" event={"ID":"ef5432f4-56a6-4518-a6b1-00ac9b29293f","Type":"ContainerStarted","Data":"5d21850cff89f6dbdb62b7a09acfce96d4cc9f631f7f8d5aaf86758b201eaeca"} Dec 06 16:07:46 crc kubenswrapper[4813]: I1206 16:07:46.537290 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2" event={"ID":"ef5432f4-56a6-4518-a6b1-00ac9b29293f","Type":"ContainerStarted","Data":"4b2d706e57865a6703adf0e7c13f3392b67a291f28fc2052189ece67a59ac18a"} Dec 06 16:07:46 crc kubenswrapper[4813]: I1206 16:07:46.568320 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2" podStartSLOduration=2.633800368 podStartE2EDuration="12.568296193s" podCreationTimestamp="2025-12-06 16:07:34 +0000 UTC" firstStartedPulling="2025-12-06 16:07:35.670567596 +0000 UTC m=+1295.561447172" lastFinishedPulling="2025-12-06 16:07:45.605063421 +0000 UTC m=+1305.495942997" observedRunningTime="2025-12-06 16:07:46.562553311 +0000 UTC m=+1306.453432897" watchObservedRunningTime="2025-12-06 16:07:46.568296193 +0000 UTC m=+1306.459175799" Dec 06 16:07:47 crc kubenswrapper[4813]: I1206 16:07:47.596360 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 06 16:07:48 crc kubenswrapper[4813]: I1206 16:07:48.634447 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 06 16:07:57 crc kubenswrapper[4813]: I1206 16:07:57.660135 4813 generic.go:334] "Generic (PLEG): container finished" podID="ef5432f4-56a6-4518-a6b1-00ac9b29293f" containerID="4b2d706e57865a6703adf0e7c13f3392b67a291f28fc2052189ece67a59ac18a" exitCode=0 Dec 06 16:07:57 crc kubenswrapper[4813]: I1206 16:07:57.660232 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2" event={"ID":"ef5432f4-56a6-4518-a6b1-00ac9b29293f","Type":"ContainerDied","Data":"4b2d706e57865a6703adf0e7c13f3392b67a291f28fc2052189ece67a59ac18a"} Dec 06 16:07:59 crc kubenswrapper[4813]: I1206 16:07:59.139577 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2" Dec 06 16:07:59 crc kubenswrapper[4813]: I1206 16:07:59.214914 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sz46d\" (UniqueName: \"kubernetes.io/projected/ef5432f4-56a6-4518-a6b1-00ac9b29293f-kube-api-access-sz46d\") pod \"ef5432f4-56a6-4518-a6b1-00ac9b29293f\" (UID: \"ef5432f4-56a6-4518-a6b1-00ac9b29293f\") " Dec 06 16:07:59 crc kubenswrapper[4813]: I1206 16:07:59.215028 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef5432f4-56a6-4518-a6b1-00ac9b29293f-repo-setup-combined-ca-bundle\") pod \"ef5432f4-56a6-4518-a6b1-00ac9b29293f\" (UID: \"ef5432f4-56a6-4518-a6b1-00ac9b29293f\") " Dec 06 16:07:59 crc kubenswrapper[4813]: I1206 16:07:59.215213 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef5432f4-56a6-4518-a6b1-00ac9b29293f-inventory\") pod \"ef5432f4-56a6-4518-a6b1-00ac9b29293f\" (UID: \"ef5432f4-56a6-4518-a6b1-00ac9b29293f\") " Dec 06 16:07:59 crc kubenswrapper[4813]: I1206 16:07:59.215249 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef5432f4-56a6-4518-a6b1-00ac9b29293f-ssh-key\") pod \"ef5432f4-56a6-4518-a6b1-00ac9b29293f\" (UID: \"ef5432f4-56a6-4518-a6b1-00ac9b29293f\") " Dec 06 16:07:59 crc kubenswrapper[4813]: I1206 16:07:59.224468 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef5432f4-56a6-4518-a6b1-00ac9b29293f-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "ef5432f4-56a6-4518-a6b1-00ac9b29293f" (UID: "ef5432f4-56a6-4518-a6b1-00ac9b29293f"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:07:59 crc kubenswrapper[4813]: I1206 16:07:59.238728 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef5432f4-56a6-4518-a6b1-00ac9b29293f-kube-api-access-sz46d" (OuterVolumeSpecName: "kube-api-access-sz46d") pod "ef5432f4-56a6-4518-a6b1-00ac9b29293f" (UID: "ef5432f4-56a6-4518-a6b1-00ac9b29293f"). InnerVolumeSpecName "kube-api-access-sz46d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:07:59 crc kubenswrapper[4813]: I1206 16:07:59.252186 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef5432f4-56a6-4518-a6b1-00ac9b29293f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ef5432f4-56a6-4518-a6b1-00ac9b29293f" (UID: "ef5432f4-56a6-4518-a6b1-00ac9b29293f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:07:59 crc kubenswrapper[4813]: I1206 16:07:59.254433 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef5432f4-56a6-4518-a6b1-00ac9b29293f-inventory" (OuterVolumeSpecName: "inventory") pod "ef5432f4-56a6-4518-a6b1-00ac9b29293f" (UID: "ef5432f4-56a6-4518-a6b1-00ac9b29293f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:07:59 crc kubenswrapper[4813]: I1206 16:07:59.317521 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sz46d\" (UniqueName: \"kubernetes.io/projected/ef5432f4-56a6-4518-a6b1-00ac9b29293f-kube-api-access-sz46d\") on node \"crc\" DevicePath \"\"" Dec 06 16:07:59 crc kubenswrapper[4813]: I1206 16:07:59.317548 4813 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef5432f4-56a6-4518-a6b1-00ac9b29293f-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:07:59 crc kubenswrapper[4813]: I1206 16:07:59.317558 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef5432f4-56a6-4518-a6b1-00ac9b29293f-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 16:07:59 crc kubenswrapper[4813]: I1206 16:07:59.317569 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef5432f4-56a6-4518-a6b1-00ac9b29293f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 16:07:59 crc kubenswrapper[4813]: I1206 16:07:59.699427 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2" event={"ID":"ef5432f4-56a6-4518-a6b1-00ac9b29293f","Type":"ContainerDied","Data":"5d21850cff89f6dbdb62b7a09acfce96d4cc9f631f7f8d5aaf86758b201eaeca"} Dec 06 16:07:59 crc kubenswrapper[4813]: I1206 16:07:59.699479 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d21850cff89f6dbdb62b7a09acfce96d4cc9f631f7f8d5aaf86758b201eaeca" Dec 06 16:07:59 crc kubenswrapper[4813]: I1206 16:07:59.699554 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2" Dec 06 16:07:59 crc kubenswrapper[4813]: I1206 16:07:59.835141 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6p6l"] Dec 06 16:07:59 crc kubenswrapper[4813]: E1206 16:07:59.835746 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef5432f4-56a6-4518-a6b1-00ac9b29293f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 06 16:07:59 crc kubenswrapper[4813]: I1206 16:07:59.835764 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef5432f4-56a6-4518-a6b1-00ac9b29293f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 06 16:07:59 crc kubenswrapper[4813]: I1206 16:07:59.835969 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef5432f4-56a6-4518-a6b1-00ac9b29293f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 06 16:07:59 crc kubenswrapper[4813]: I1206 16:07:59.836554 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6p6l" Dec 06 16:07:59 crc kubenswrapper[4813]: I1206 16:07:59.841662 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 16:07:59 crc kubenswrapper[4813]: I1206 16:07:59.842751 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 16:07:59 crc kubenswrapper[4813]: I1206 16:07:59.842799 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jvkhf" Dec 06 16:07:59 crc kubenswrapper[4813]: I1206 16:07:59.845099 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 16:07:59 crc kubenswrapper[4813]: I1206 16:07:59.865765 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6p6l"] Dec 06 16:07:59 crc kubenswrapper[4813]: I1206 16:07:59.928882 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5mr6\" (UniqueName: \"kubernetes.io/projected/aa40ab8b-bc66-4e51-a126-ec15a6250486-kube-api-access-z5mr6\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-n6p6l\" (UID: \"aa40ab8b-bc66-4e51-a126-ec15a6250486\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6p6l" Dec 06 16:07:59 crc kubenswrapper[4813]: I1206 16:07:59.929010 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa40ab8b-bc66-4e51-a126-ec15a6250486-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-n6p6l\" (UID: \"aa40ab8b-bc66-4e51-a126-ec15a6250486\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6p6l" Dec 06 16:07:59 crc kubenswrapper[4813]: I1206 16:07:59.929058 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa40ab8b-bc66-4e51-a126-ec15a6250486-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-n6p6l\" (UID: \"aa40ab8b-bc66-4e51-a126-ec15a6250486\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6p6l" Dec 06 16:08:00 crc kubenswrapper[4813]: I1206 16:08:00.030597 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5mr6\" (UniqueName: \"kubernetes.io/projected/aa40ab8b-bc66-4e51-a126-ec15a6250486-kube-api-access-z5mr6\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-n6p6l\" (UID: \"aa40ab8b-bc66-4e51-a126-ec15a6250486\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6p6l" Dec 06 16:08:00 crc kubenswrapper[4813]: I1206 16:08:00.030712 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa40ab8b-bc66-4e51-a126-ec15a6250486-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-n6p6l\" (UID: \"aa40ab8b-bc66-4e51-a126-ec15a6250486\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6p6l" Dec 06 16:08:00 crc kubenswrapper[4813]: I1206 16:08:00.030759 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa40ab8b-bc66-4e51-a126-ec15a6250486-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-n6p6l\" (UID: \"aa40ab8b-bc66-4e51-a126-ec15a6250486\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6p6l" Dec 06 16:08:00 crc kubenswrapper[4813]: I1206 16:08:00.035812 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa40ab8b-bc66-4e51-a126-ec15a6250486-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-n6p6l\" (UID: \"aa40ab8b-bc66-4e51-a126-ec15a6250486\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6p6l" Dec 06 16:08:00 crc kubenswrapper[4813]: I1206 16:08:00.047243 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa40ab8b-bc66-4e51-a126-ec15a6250486-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-n6p6l\" (UID: \"aa40ab8b-bc66-4e51-a126-ec15a6250486\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6p6l" Dec 06 16:08:00 crc kubenswrapper[4813]: I1206 16:08:00.048320 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5mr6\" (UniqueName: \"kubernetes.io/projected/aa40ab8b-bc66-4e51-a126-ec15a6250486-kube-api-access-z5mr6\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-n6p6l\" (UID: \"aa40ab8b-bc66-4e51-a126-ec15a6250486\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6p6l" Dec 06 16:08:00 crc kubenswrapper[4813]: I1206 16:08:00.152241 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6p6l" Dec 06 16:08:00 crc kubenswrapper[4813]: I1206 16:08:00.765555 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6p6l"] Dec 06 16:08:01 crc kubenswrapper[4813]: I1206 16:08:01.173784 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 16:08:01 crc kubenswrapper[4813]: I1206 16:08:01.719993 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6p6l" event={"ID":"aa40ab8b-bc66-4e51-a126-ec15a6250486","Type":"ContainerStarted","Data":"ee02a09ba5385e32be34f4c7fb70c278a02eaaec7f005801870f8d1080b8f24b"} Dec 06 16:08:01 crc kubenswrapper[4813]: I1206 16:08:01.720253 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6p6l" event={"ID":"aa40ab8b-bc66-4e51-a126-ec15a6250486","Type":"ContainerStarted","Data":"fdd86999cf799326f2e9cabde8886270e4ef1a559ec23d17c355e3f5d3fe32bb"} Dec 06 16:08:01 crc kubenswrapper[4813]: I1206 16:08:01.731454 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6p6l" podStartSLOduration=2.335813862 podStartE2EDuration="2.731437263s" podCreationTimestamp="2025-12-06 16:07:59 +0000 UTC" firstStartedPulling="2025-12-06 16:08:00.775391612 +0000 UTC m=+1320.666271208" lastFinishedPulling="2025-12-06 16:08:01.171015033 +0000 UTC m=+1321.061894609" observedRunningTime="2025-12-06 16:08:01.731212347 +0000 UTC m=+1321.622091933" watchObservedRunningTime="2025-12-06 16:08:01.731437263 +0000 UTC m=+1321.622316839" Dec 06 16:08:04 crc kubenswrapper[4813]: I1206 16:08:04.753737 4813 generic.go:334] "Generic (PLEG): container finished" podID="aa40ab8b-bc66-4e51-a126-ec15a6250486" containerID="ee02a09ba5385e32be34f4c7fb70c278a02eaaec7f005801870f8d1080b8f24b" exitCode=0 Dec 06 16:08:04 crc kubenswrapper[4813]: I1206 16:08:04.753946 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6p6l" event={"ID":"aa40ab8b-bc66-4e51-a126-ec15a6250486","Type":"ContainerDied","Data":"ee02a09ba5385e32be34f4c7fb70c278a02eaaec7f005801870f8d1080b8f24b"} Dec 06 16:08:06 crc kubenswrapper[4813]: I1206 16:08:06.314386 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6p6l" Dec 06 16:08:06 crc kubenswrapper[4813]: I1206 16:08:06.353580 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5mr6\" (UniqueName: \"kubernetes.io/projected/aa40ab8b-bc66-4e51-a126-ec15a6250486-kube-api-access-z5mr6\") pod \"aa40ab8b-bc66-4e51-a126-ec15a6250486\" (UID: \"aa40ab8b-bc66-4e51-a126-ec15a6250486\") " Dec 06 16:08:06 crc kubenswrapper[4813]: I1206 16:08:06.353906 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa40ab8b-bc66-4e51-a126-ec15a6250486-ssh-key\") pod \"aa40ab8b-bc66-4e51-a126-ec15a6250486\" (UID: \"aa40ab8b-bc66-4e51-a126-ec15a6250486\") " Dec 06 16:08:06 crc kubenswrapper[4813]: I1206 16:08:06.354157 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa40ab8b-bc66-4e51-a126-ec15a6250486-inventory\") pod \"aa40ab8b-bc66-4e51-a126-ec15a6250486\" (UID: \"aa40ab8b-bc66-4e51-a126-ec15a6250486\") " Dec 06 16:08:06 crc kubenswrapper[4813]: I1206 16:08:06.359533 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa40ab8b-bc66-4e51-a126-ec15a6250486-kube-api-access-z5mr6" (OuterVolumeSpecName: "kube-api-access-z5mr6") pod "aa40ab8b-bc66-4e51-a126-ec15a6250486" (UID: "aa40ab8b-bc66-4e51-a126-ec15a6250486"). InnerVolumeSpecName "kube-api-access-z5mr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:08:06 crc kubenswrapper[4813]: I1206 16:08:06.383288 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa40ab8b-bc66-4e51-a126-ec15a6250486-inventory" (OuterVolumeSpecName: "inventory") pod "aa40ab8b-bc66-4e51-a126-ec15a6250486" (UID: "aa40ab8b-bc66-4e51-a126-ec15a6250486"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:08:06 crc kubenswrapper[4813]: I1206 16:08:06.383933 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa40ab8b-bc66-4e51-a126-ec15a6250486-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "aa40ab8b-bc66-4e51-a126-ec15a6250486" (UID: "aa40ab8b-bc66-4e51-a126-ec15a6250486"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:08:06 crc kubenswrapper[4813]: I1206 16:08:06.455664 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5mr6\" (UniqueName: \"kubernetes.io/projected/aa40ab8b-bc66-4e51-a126-ec15a6250486-kube-api-access-z5mr6\") on node \"crc\" DevicePath \"\"" Dec 06 16:08:06 crc kubenswrapper[4813]: I1206 16:08:06.455689 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa40ab8b-bc66-4e51-a126-ec15a6250486-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 16:08:06 crc kubenswrapper[4813]: I1206 16:08:06.455698 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa40ab8b-bc66-4e51-a126-ec15a6250486-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 16:08:06 crc kubenswrapper[4813]: I1206 16:08:06.783129 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6p6l" event={"ID":"aa40ab8b-bc66-4e51-a126-ec15a6250486","Type":"ContainerDied","Data":"fdd86999cf799326f2e9cabde8886270e4ef1a559ec23d17c355e3f5d3fe32bb"} Dec 06 16:08:06 crc kubenswrapper[4813]: I1206 16:08:06.783460 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdd86999cf799326f2e9cabde8886270e4ef1a559ec23d17c355e3f5d3fe32bb" Dec 06 16:08:06 crc kubenswrapper[4813]: I1206 16:08:06.783404 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6p6l" Dec 06 16:08:06 crc kubenswrapper[4813]: I1206 16:08:06.878664 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qr892"] Dec 06 16:08:06 crc kubenswrapper[4813]: E1206 16:08:06.879029 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa40ab8b-bc66-4e51-a126-ec15a6250486" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 06 16:08:06 crc kubenswrapper[4813]: I1206 16:08:06.879045 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa40ab8b-bc66-4e51-a126-ec15a6250486" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 06 16:08:06 crc kubenswrapper[4813]: I1206 16:08:06.879211 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa40ab8b-bc66-4e51-a126-ec15a6250486" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 06 16:08:06 crc kubenswrapper[4813]: I1206 16:08:06.879788 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qr892" Dec 06 16:08:06 crc kubenswrapper[4813]: I1206 16:08:06.882440 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 16:08:06 crc kubenswrapper[4813]: I1206 16:08:06.882855 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 16:08:06 crc kubenswrapper[4813]: I1206 16:08:06.883508 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 16:08:06 crc kubenswrapper[4813]: I1206 16:08:06.887400 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jvkhf" Dec 06 16:08:06 crc kubenswrapper[4813]: I1206 16:08:06.948666 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qr892"] Dec 06 16:08:06 crc kubenswrapper[4813]: I1206 16:08:06.968951 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6306b1cf-bf21-4cd0-a7c8-c83a99863266-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qr892\" (UID: \"6306b1cf-bf21-4cd0-a7c8-c83a99863266\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qr892" Dec 06 16:08:06 crc kubenswrapper[4813]: I1206 16:08:06.969060 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55l5d\" (UniqueName: \"kubernetes.io/projected/6306b1cf-bf21-4cd0-a7c8-c83a99863266-kube-api-access-55l5d\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qr892\" (UID: \"6306b1cf-bf21-4cd0-a7c8-c83a99863266\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qr892" Dec 06 16:08:06 crc kubenswrapper[4813]: I1206 16:08:06.969115 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6306b1cf-bf21-4cd0-a7c8-c83a99863266-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qr892\" (UID: \"6306b1cf-bf21-4cd0-a7c8-c83a99863266\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qr892" Dec 06 16:08:06 crc kubenswrapper[4813]: I1206 16:08:06.969478 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6306b1cf-bf21-4cd0-a7c8-c83a99863266-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qr892\" (UID: \"6306b1cf-bf21-4cd0-a7c8-c83a99863266\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qr892" Dec 06 16:08:07 crc kubenswrapper[4813]: I1206 16:08:07.071030 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6306b1cf-bf21-4cd0-a7c8-c83a99863266-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qr892\" (UID: \"6306b1cf-bf21-4cd0-a7c8-c83a99863266\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qr892" Dec 06 16:08:07 crc kubenswrapper[4813]: I1206 16:08:07.071418 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6306b1cf-bf21-4cd0-a7c8-c83a99863266-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qr892\" (UID: \"6306b1cf-bf21-4cd0-a7c8-c83a99863266\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qr892" Dec 06 16:08:07 crc kubenswrapper[4813]: I1206 16:08:07.071600 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55l5d\" (UniqueName: \"kubernetes.io/projected/6306b1cf-bf21-4cd0-a7c8-c83a99863266-kube-api-access-55l5d\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qr892\" (UID: \"6306b1cf-bf21-4cd0-a7c8-c83a99863266\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qr892" Dec 06 16:08:07 crc kubenswrapper[4813]: I1206 16:08:07.071718 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6306b1cf-bf21-4cd0-a7c8-c83a99863266-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qr892\" (UID: \"6306b1cf-bf21-4cd0-a7c8-c83a99863266\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qr892" Dec 06 16:08:07 crc kubenswrapper[4813]: I1206 16:08:07.076321 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6306b1cf-bf21-4cd0-a7c8-c83a99863266-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qr892\" (UID: \"6306b1cf-bf21-4cd0-a7c8-c83a99863266\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qr892" Dec 06 16:08:07 crc kubenswrapper[4813]: I1206 16:08:07.076477 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6306b1cf-bf21-4cd0-a7c8-c83a99863266-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qr892\" (UID: \"6306b1cf-bf21-4cd0-a7c8-c83a99863266\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qr892" Dec 06 16:08:07 crc kubenswrapper[4813]: I1206 16:08:07.076696 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6306b1cf-bf21-4cd0-a7c8-c83a99863266-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qr892\" (UID: \"6306b1cf-bf21-4cd0-a7c8-c83a99863266\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qr892" Dec 06 16:08:07 crc kubenswrapper[4813]: I1206 16:08:07.094422 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55l5d\" (UniqueName: \"kubernetes.io/projected/6306b1cf-bf21-4cd0-a7c8-c83a99863266-kube-api-access-55l5d\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qr892\" (UID: \"6306b1cf-bf21-4cd0-a7c8-c83a99863266\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qr892" Dec 06 16:08:07 crc kubenswrapper[4813]: I1206 16:08:07.200373 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qr892" Dec 06 16:08:07 crc kubenswrapper[4813]: I1206 16:08:07.803173 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qr892"] Dec 06 16:08:08 crc kubenswrapper[4813]: I1206 16:08:08.807810 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qr892" event={"ID":"6306b1cf-bf21-4cd0-a7c8-c83a99863266","Type":"ContainerStarted","Data":"b56853c7861661f6a657ef85062b0040b736154e0fd63d5d7a7e3f0ccd86103c"} Dec 06 16:08:08 crc kubenswrapper[4813]: I1206 16:08:08.808118 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qr892" event={"ID":"6306b1cf-bf21-4cd0-a7c8-c83a99863266","Type":"ContainerStarted","Data":"26b732f26368a1b87e53854bec4306c340399e83779aebe0eb45bbbd1c4b3064"} Dec 06 16:08:08 crc kubenswrapper[4813]: I1206 16:08:08.843312 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qr892" podStartSLOduration=2.432542277 podStartE2EDuration="2.843286857s" podCreationTimestamp="2025-12-06 16:08:06 +0000 UTC" firstStartedPulling="2025-12-06 16:08:07.81210757 +0000 UTC m=+1327.702987166" lastFinishedPulling="2025-12-06 16:08:08.22285214 +0000 UTC m=+1328.113731746" observedRunningTime="2025-12-06 16:08:08.831497305 +0000 UTC m=+1328.722376891" watchObservedRunningTime="2025-12-06 16:08:08.843286857 +0000 UTC m=+1328.734166443" Dec 06 16:08:19 crc kubenswrapper[4813]: I1206 16:08:19.428631 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:08:19 crc kubenswrapper[4813]: I1206 16:08:19.429302 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:08:22 crc kubenswrapper[4813]: I1206 16:08:22.430893 4813 scope.go:117] "RemoveContainer" containerID="40897f9b7d87e956550d6ad37725608ed528e8b32b221fdab9ff83f5da018545" Dec 06 16:08:49 crc kubenswrapper[4813]: I1206 16:08:49.427523 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:08:49 crc kubenswrapper[4813]: I1206 16:08:49.428265 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:08:54 crc kubenswrapper[4813]: I1206 16:08:54.956908 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="bf7377f8-ef23-484a-b66b-90db1cd5b7cd" containerName="galera" probeResult="failure" output="command timed out" Dec 06 16:08:54 crc kubenswrapper[4813]: I1206 16:08:54.957800 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="bf7377f8-ef23-484a-b66b-90db1cd5b7cd" containerName="galera" probeResult="failure" output="command timed out" Dec 06 16:09:19 crc kubenswrapper[4813]: I1206 16:09:19.428266 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:09:19 crc kubenswrapper[4813]: I1206 16:09:19.428988 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:09:19 crc kubenswrapper[4813]: I1206 16:09:19.429051 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" Dec 06 16:09:19 crc kubenswrapper[4813]: I1206 16:09:19.429917 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f25008e92574e2c832578c1ff4d4518d3c2eb82b415ff9e3dd99f9504d835fc6"} pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 16:09:19 crc kubenswrapper[4813]: I1206 16:09:19.430011 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" containerID="cri-o://f25008e92574e2c832578c1ff4d4518d3c2eb82b415ff9e3dd99f9504d835fc6" gracePeriod=600 Dec 06 16:09:19 crc kubenswrapper[4813]: I1206 16:09:19.648299 4813 generic.go:334] "Generic (PLEG): container finished" podID="d88e8bae-c055-4c55-b548-f621ff96de06" containerID="f25008e92574e2c832578c1ff4d4518d3c2eb82b415ff9e3dd99f9504d835fc6" exitCode=0 Dec 06 16:09:19 crc kubenswrapper[4813]: I1206 16:09:19.648346 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerDied","Data":"f25008e92574e2c832578c1ff4d4518d3c2eb82b415ff9e3dd99f9504d835fc6"} Dec 06 16:09:19 crc kubenswrapper[4813]: I1206 16:09:19.648390 4813 scope.go:117] "RemoveContainer" containerID="b7b68c16444a3326ba33662bc5c9583c0b689e2b31a8a7643be0512e87a18c96" Dec 06 16:09:20 crc kubenswrapper[4813]: I1206 16:09:20.660337 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerStarted","Data":"dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412"} Dec 06 16:09:22 crc kubenswrapper[4813]: I1206 16:09:22.516412 4813 scope.go:117] "RemoveContainer" containerID="bddf9da2a88bc9810142b3c40b3de2ff32912f6eb820812c93b8fa810c085281" Dec 06 16:11:03 crc kubenswrapper[4813]: I1206 16:11:03.329455 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2tr6m"] Dec 06 16:11:03 crc kubenswrapper[4813]: I1206 16:11:03.334964 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2tr6m" Dec 06 16:11:03 crc kubenswrapper[4813]: I1206 16:11:03.367152 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2tr6m"] Dec 06 16:11:03 crc kubenswrapper[4813]: I1206 16:11:03.376059 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca8aaef0-5e4f-4137-aa1c-efaccd341564-catalog-content\") pod \"redhat-marketplace-2tr6m\" (UID: \"ca8aaef0-5e4f-4137-aa1c-efaccd341564\") " pod="openshift-marketplace/redhat-marketplace-2tr6m" Dec 06 16:11:03 crc kubenswrapper[4813]: I1206 16:11:03.376166 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wh9mt\" (UniqueName: \"kubernetes.io/projected/ca8aaef0-5e4f-4137-aa1c-efaccd341564-kube-api-access-wh9mt\") pod \"redhat-marketplace-2tr6m\" (UID: \"ca8aaef0-5e4f-4137-aa1c-efaccd341564\") " pod="openshift-marketplace/redhat-marketplace-2tr6m" Dec 06 16:11:03 crc kubenswrapper[4813]: I1206 16:11:03.376244 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca8aaef0-5e4f-4137-aa1c-efaccd341564-utilities\") pod \"redhat-marketplace-2tr6m\" (UID: \"ca8aaef0-5e4f-4137-aa1c-efaccd341564\") " pod="openshift-marketplace/redhat-marketplace-2tr6m" Dec 06 16:11:03 crc kubenswrapper[4813]: I1206 16:11:03.477655 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca8aaef0-5e4f-4137-aa1c-efaccd341564-catalog-content\") pod \"redhat-marketplace-2tr6m\" (UID: \"ca8aaef0-5e4f-4137-aa1c-efaccd341564\") " pod="openshift-marketplace/redhat-marketplace-2tr6m" Dec 06 16:11:03 crc kubenswrapper[4813]: I1206 16:11:03.477758 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wh9mt\" (UniqueName: \"kubernetes.io/projected/ca8aaef0-5e4f-4137-aa1c-efaccd341564-kube-api-access-wh9mt\") pod \"redhat-marketplace-2tr6m\" (UID: \"ca8aaef0-5e4f-4137-aa1c-efaccd341564\") " pod="openshift-marketplace/redhat-marketplace-2tr6m" Dec 06 16:11:03 crc kubenswrapper[4813]: I1206 16:11:03.477795 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca8aaef0-5e4f-4137-aa1c-efaccd341564-utilities\") pod \"redhat-marketplace-2tr6m\" (UID: \"ca8aaef0-5e4f-4137-aa1c-efaccd341564\") " pod="openshift-marketplace/redhat-marketplace-2tr6m" Dec 06 16:11:03 crc kubenswrapper[4813]: I1206 16:11:03.478171 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca8aaef0-5e4f-4137-aa1c-efaccd341564-catalog-content\") pod \"redhat-marketplace-2tr6m\" (UID: \"ca8aaef0-5e4f-4137-aa1c-efaccd341564\") " pod="openshift-marketplace/redhat-marketplace-2tr6m" Dec 06 16:11:03 crc kubenswrapper[4813]: I1206 16:11:03.478206 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca8aaef0-5e4f-4137-aa1c-efaccd341564-utilities\") pod \"redhat-marketplace-2tr6m\" (UID: \"ca8aaef0-5e4f-4137-aa1c-efaccd341564\") " pod="openshift-marketplace/redhat-marketplace-2tr6m" Dec 06 16:11:03 crc kubenswrapper[4813]: I1206 16:11:03.513389 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wh9mt\" (UniqueName: \"kubernetes.io/projected/ca8aaef0-5e4f-4137-aa1c-efaccd341564-kube-api-access-wh9mt\") pod \"redhat-marketplace-2tr6m\" (UID: \"ca8aaef0-5e4f-4137-aa1c-efaccd341564\") " pod="openshift-marketplace/redhat-marketplace-2tr6m" Dec 06 16:11:03 crc kubenswrapper[4813]: I1206 16:11:03.661035 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2tr6m" Dec 06 16:11:04 crc kubenswrapper[4813]: I1206 16:11:04.119801 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2tr6m"] Dec 06 16:11:04 crc kubenswrapper[4813]: I1206 16:11:04.911979 4813 generic.go:334] "Generic (PLEG): container finished" podID="ca8aaef0-5e4f-4137-aa1c-efaccd341564" containerID="13be1f6435dc1493a19ba9791d354ac1a702cb7e3ff62b36ea3e1faf7a6d4b0a" exitCode=0 Dec 06 16:11:04 crc kubenswrapper[4813]: I1206 16:11:04.912164 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2tr6m" event={"ID":"ca8aaef0-5e4f-4137-aa1c-efaccd341564","Type":"ContainerDied","Data":"13be1f6435dc1493a19ba9791d354ac1a702cb7e3ff62b36ea3e1faf7a6d4b0a"} Dec 06 16:11:04 crc kubenswrapper[4813]: I1206 16:11:04.916377 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2tr6m" event={"ID":"ca8aaef0-5e4f-4137-aa1c-efaccd341564","Type":"ContainerStarted","Data":"2bf022f5b7e7da3343e2899f6b86f013f9dfd6eb7db8fbb2a4c371e96e5a27b8"} Dec 06 16:11:04 crc kubenswrapper[4813]: I1206 16:11:04.914139 4813 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 16:11:05 crc kubenswrapper[4813]: I1206 16:11:05.928470 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2tr6m" event={"ID":"ca8aaef0-5e4f-4137-aa1c-efaccd341564","Type":"ContainerStarted","Data":"47983e4b2a3a963ff3a42167aed0e8dc2dbf4dc65714b586c7f2592fd42d1a7f"} Dec 06 16:11:06 crc kubenswrapper[4813]: I1206 16:11:06.945485 4813 generic.go:334] "Generic (PLEG): container finished" podID="ca8aaef0-5e4f-4137-aa1c-efaccd341564" containerID="47983e4b2a3a963ff3a42167aed0e8dc2dbf4dc65714b586c7f2592fd42d1a7f" exitCode=0 Dec 06 16:11:06 crc kubenswrapper[4813]: I1206 16:11:06.945651 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2tr6m" event={"ID":"ca8aaef0-5e4f-4137-aa1c-efaccd341564","Type":"ContainerDied","Data":"47983e4b2a3a963ff3a42167aed0e8dc2dbf4dc65714b586c7f2592fd42d1a7f"} Dec 06 16:11:07 crc kubenswrapper[4813]: I1206 16:11:07.965346 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2tr6m" event={"ID":"ca8aaef0-5e4f-4137-aa1c-efaccd341564","Type":"ContainerStarted","Data":"b5f42bfa3d0849de967b8eb10a1f4d5dc475be58e2c04c102e0f29e914b1eb83"} Dec 06 16:11:13 crc kubenswrapper[4813]: I1206 16:11:13.661789 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2tr6m" Dec 06 16:11:13 crc kubenswrapper[4813]: I1206 16:11:13.662348 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2tr6m" Dec 06 16:11:13 crc kubenswrapper[4813]: I1206 16:11:13.714241 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2tr6m" Dec 06 16:11:13 crc kubenswrapper[4813]: I1206 16:11:13.766490 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2tr6m" podStartSLOduration=8.293004404 podStartE2EDuration="10.766468341s" podCreationTimestamp="2025-12-06 16:11:03 +0000 UTC" firstStartedPulling="2025-12-06 16:11:04.913879661 +0000 UTC m=+1504.804759237" lastFinishedPulling="2025-12-06 16:11:07.387343568 +0000 UTC m=+1507.278223174" observedRunningTime="2025-12-06 16:11:07.994875812 +0000 UTC m=+1507.885755388" watchObservedRunningTime="2025-12-06 16:11:13.766468341 +0000 UTC m=+1513.657347917" Dec 06 16:11:14 crc kubenswrapper[4813]: I1206 16:11:14.189235 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2tr6m" Dec 06 16:11:14 crc kubenswrapper[4813]: I1206 16:11:14.245514 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2tr6m"] Dec 06 16:11:16 crc kubenswrapper[4813]: I1206 16:11:16.060078 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2tr6m" podUID="ca8aaef0-5e4f-4137-aa1c-efaccd341564" containerName="registry-server" containerID="cri-o://b5f42bfa3d0849de967b8eb10a1f4d5dc475be58e2c04c102e0f29e914b1eb83" gracePeriod=2 Dec 06 16:11:16 crc kubenswrapper[4813]: I1206 16:11:16.558514 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2tr6m" Dec 06 16:11:16 crc kubenswrapper[4813]: I1206 16:11:16.669735 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wh9mt\" (UniqueName: \"kubernetes.io/projected/ca8aaef0-5e4f-4137-aa1c-efaccd341564-kube-api-access-wh9mt\") pod \"ca8aaef0-5e4f-4137-aa1c-efaccd341564\" (UID: \"ca8aaef0-5e4f-4137-aa1c-efaccd341564\") " Dec 06 16:11:16 crc kubenswrapper[4813]: I1206 16:11:16.669895 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca8aaef0-5e4f-4137-aa1c-efaccd341564-catalog-content\") pod \"ca8aaef0-5e4f-4137-aa1c-efaccd341564\" (UID: \"ca8aaef0-5e4f-4137-aa1c-efaccd341564\") " Dec 06 16:11:16 crc kubenswrapper[4813]: I1206 16:11:16.669977 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca8aaef0-5e4f-4137-aa1c-efaccd341564-utilities\") pod \"ca8aaef0-5e4f-4137-aa1c-efaccd341564\" (UID: \"ca8aaef0-5e4f-4137-aa1c-efaccd341564\") " Dec 06 16:11:16 crc kubenswrapper[4813]: I1206 16:11:16.670621 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca8aaef0-5e4f-4137-aa1c-efaccd341564-utilities" (OuterVolumeSpecName: "utilities") pod "ca8aaef0-5e4f-4137-aa1c-efaccd341564" (UID: "ca8aaef0-5e4f-4137-aa1c-efaccd341564"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:11:16 crc kubenswrapper[4813]: I1206 16:11:16.689544 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca8aaef0-5e4f-4137-aa1c-efaccd341564-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ca8aaef0-5e4f-4137-aa1c-efaccd341564" (UID: "ca8aaef0-5e4f-4137-aa1c-efaccd341564"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:11:16 crc kubenswrapper[4813]: I1206 16:11:16.689666 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca8aaef0-5e4f-4137-aa1c-efaccd341564-kube-api-access-wh9mt" (OuterVolumeSpecName: "kube-api-access-wh9mt") pod "ca8aaef0-5e4f-4137-aa1c-efaccd341564" (UID: "ca8aaef0-5e4f-4137-aa1c-efaccd341564"). InnerVolumeSpecName "kube-api-access-wh9mt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:11:16 crc kubenswrapper[4813]: I1206 16:11:16.772467 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca8aaef0-5e4f-4137-aa1c-efaccd341564-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 16:11:16 crc kubenswrapper[4813]: I1206 16:11:16.772498 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wh9mt\" (UniqueName: \"kubernetes.io/projected/ca8aaef0-5e4f-4137-aa1c-efaccd341564-kube-api-access-wh9mt\") on node \"crc\" DevicePath \"\"" Dec 06 16:11:16 crc kubenswrapper[4813]: I1206 16:11:16.772511 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca8aaef0-5e4f-4137-aa1c-efaccd341564-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 16:11:17 crc kubenswrapper[4813]: I1206 16:11:17.069542 4813 generic.go:334] "Generic (PLEG): container finished" podID="ca8aaef0-5e4f-4137-aa1c-efaccd341564" containerID="b5f42bfa3d0849de967b8eb10a1f4d5dc475be58e2c04c102e0f29e914b1eb83" exitCode=0 Dec 06 16:11:17 crc kubenswrapper[4813]: I1206 16:11:17.069604 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2tr6m" Dec 06 16:11:17 crc kubenswrapper[4813]: I1206 16:11:17.069630 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2tr6m" event={"ID":"ca8aaef0-5e4f-4137-aa1c-efaccd341564","Type":"ContainerDied","Data":"b5f42bfa3d0849de967b8eb10a1f4d5dc475be58e2c04c102e0f29e914b1eb83"} Dec 06 16:11:17 crc kubenswrapper[4813]: I1206 16:11:17.070327 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2tr6m" event={"ID":"ca8aaef0-5e4f-4137-aa1c-efaccd341564","Type":"ContainerDied","Data":"2bf022f5b7e7da3343e2899f6b86f013f9dfd6eb7db8fbb2a4c371e96e5a27b8"} Dec 06 16:11:17 crc kubenswrapper[4813]: I1206 16:11:17.070364 4813 scope.go:117] "RemoveContainer" containerID="b5f42bfa3d0849de967b8eb10a1f4d5dc475be58e2c04c102e0f29e914b1eb83" Dec 06 16:11:17 crc kubenswrapper[4813]: I1206 16:11:17.090446 4813 scope.go:117] "RemoveContainer" containerID="47983e4b2a3a963ff3a42167aed0e8dc2dbf4dc65714b586c7f2592fd42d1a7f" Dec 06 16:11:17 crc kubenswrapper[4813]: I1206 16:11:17.105137 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2tr6m"] Dec 06 16:11:17 crc kubenswrapper[4813]: I1206 16:11:17.115346 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2tr6m"] Dec 06 16:11:17 crc kubenswrapper[4813]: I1206 16:11:17.121300 4813 scope.go:117] "RemoveContainer" containerID="13be1f6435dc1493a19ba9791d354ac1a702cb7e3ff62b36ea3e1faf7a6d4b0a" Dec 06 16:11:17 crc kubenswrapper[4813]: I1206 16:11:17.149091 4813 scope.go:117] "RemoveContainer" containerID="b5f42bfa3d0849de967b8eb10a1f4d5dc475be58e2c04c102e0f29e914b1eb83" Dec 06 16:11:17 crc kubenswrapper[4813]: E1206 16:11:17.149812 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5f42bfa3d0849de967b8eb10a1f4d5dc475be58e2c04c102e0f29e914b1eb83\": container with ID starting with b5f42bfa3d0849de967b8eb10a1f4d5dc475be58e2c04c102e0f29e914b1eb83 not found: ID does not exist" containerID="b5f42bfa3d0849de967b8eb10a1f4d5dc475be58e2c04c102e0f29e914b1eb83" Dec 06 16:11:17 crc kubenswrapper[4813]: I1206 16:11:17.149851 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5f42bfa3d0849de967b8eb10a1f4d5dc475be58e2c04c102e0f29e914b1eb83"} err="failed to get container status \"b5f42bfa3d0849de967b8eb10a1f4d5dc475be58e2c04c102e0f29e914b1eb83\": rpc error: code = NotFound desc = could not find container \"b5f42bfa3d0849de967b8eb10a1f4d5dc475be58e2c04c102e0f29e914b1eb83\": container with ID starting with b5f42bfa3d0849de967b8eb10a1f4d5dc475be58e2c04c102e0f29e914b1eb83 not found: ID does not exist" Dec 06 16:11:17 crc kubenswrapper[4813]: I1206 16:11:17.149871 4813 scope.go:117] "RemoveContainer" containerID="47983e4b2a3a963ff3a42167aed0e8dc2dbf4dc65714b586c7f2592fd42d1a7f" Dec 06 16:11:17 crc kubenswrapper[4813]: E1206 16:11:17.150232 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47983e4b2a3a963ff3a42167aed0e8dc2dbf4dc65714b586c7f2592fd42d1a7f\": container with ID starting with 47983e4b2a3a963ff3a42167aed0e8dc2dbf4dc65714b586c7f2592fd42d1a7f not found: ID does not exist" containerID="47983e4b2a3a963ff3a42167aed0e8dc2dbf4dc65714b586c7f2592fd42d1a7f" Dec 06 16:11:17 crc kubenswrapper[4813]: I1206 16:11:17.150253 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47983e4b2a3a963ff3a42167aed0e8dc2dbf4dc65714b586c7f2592fd42d1a7f"} err="failed to get container status \"47983e4b2a3a963ff3a42167aed0e8dc2dbf4dc65714b586c7f2592fd42d1a7f\": rpc error: code = NotFound desc = could not find container \"47983e4b2a3a963ff3a42167aed0e8dc2dbf4dc65714b586c7f2592fd42d1a7f\": container with ID starting with 47983e4b2a3a963ff3a42167aed0e8dc2dbf4dc65714b586c7f2592fd42d1a7f not found: ID does not exist" Dec 06 16:11:17 crc kubenswrapper[4813]: I1206 16:11:17.150278 4813 scope.go:117] "RemoveContainer" containerID="13be1f6435dc1493a19ba9791d354ac1a702cb7e3ff62b36ea3e1faf7a6d4b0a" Dec 06 16:11:17 crc kubenswrapper[4813]: E1206 16:11:17.150608 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13be1f6435dc1493a19ba9791d354ac1a702cb7e3ff62b36ea3e1faf7a6d4b0a\": container with ID starting with 13be1f6435dc1493a19ba9791d354ac1a702cb7e3ff62b36ea3e1faf7a6d4b0a not found: ID does not exist" containerID="13be1f6435dc1493a19ba9791d354ac1a702cb7e3ff62b36ea3e1faf7a6d4b0a" Dec 06 16:11:17 crc kubenswrapper[4813]: I1206 16:11:17.150650 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13be1f6435dc1493a19ba9791d354ac1a702cb7e3ff62b36ea3e1faf7a6d4b0a"} err="failed to get container status \"13be1f6435dc1493a19ba9791d354ac1a702cb7e3ff62b36ea3e1faf7a6d4b0a\": rpc error: code = NotFound desc = could not find container \"13be1f6435dc1493a19ba9791d354ac1a702cb7e3ff62b36ea3e1faf7a6d4b0a\": container with ID starting with 13be1f6435dc1493a19ba9791d354ac1a702cb7e3ff62b36ea3e1faf7a6d4b0a not found: ID does not exist" Dec 06 16:11:18 crc kubenswrapper[4813]: I1206 16:11:18.506399 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca8aaef0-5e4f-4137-aa1c-efaccd341564" path="/var/lib/kubelet/pods/ca8aaef0-5e4f-4137-aa1c-efaccd341564/volumes" Dec 06 16:11:19 crc kubenswrapper[4813]: I1206 16:11:19.402174 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zfb4t"] Dec 06 16:11:19 crc kubenswrapper[4813]: E1206 16:11:19.403003 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca8aaef0-5e4f-4137-aa1c-efaccd341564" containerName="extract-content" Dec 06 16:11:19 crc kubenswrapper[4813]: I1206 16:11:19.403027 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca8aaef0-5e4f-4137-aa1c-efaccd341564" containerName="extract-content" Dec 06 16:11:19 crc kubenswrapper[4813]: E1206 16:11:19.403042 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca8aaef0-5e4f-4137-aa1c-efaccd341564" containerName="registry-server" Dec 06 16:11:19 crc kubenswrapper[4813]: I1206 16:11:19.403050 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca8aaef0-5e4f-4137-aa1c-efaccd341564" containerName="registry-server" Dec 06 16:11:19 crc kubenswrapper[4813]: E1206 16:11:19.403099 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca8aaef0-5e4f-4137-aa1c-efaccd341564" containerName="extract-utilities" Dec 06 16:11:19 crc kubenswrapper[4813]: I1206 16:11:19.403108 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca8aaef0-5e4f-4137-aa1c-efaccd341564" containerName="extract-utilities" Dec 06 16:11:19 crc kubenswrapper[4813]: I1206 16:11:19.403356 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca8aaef0-5e4f-4137-aa1c-efaccd341564" containerName="registry-server" Dec 06 16:11:19 crc kubenswrapper[4813]: I1206 16:11:19.407601 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zfb4t" Dec 06 16:11:19 crc kubenswrapper[4813]: I1206 16:11:19.424288 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zfb4t"] Dec 06 16:11:19 crc kubenswrapper[4813]: I1206 16:11:19.428771 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:11:19 crc kubenswrapper[4813]: I1206 16:11:19.428833 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:11:19 crc kubenswrapper[4813]: I1206 16:11:19.521980 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cae6355-3939-4aac-9592-3d916f54c6bf-catalog-content\") pod \"community-operators-zfb4t\" (UID: \"3cae6355-3939-4aac-9592-3d916f54c6bf\") " pod="openshift-marketplace/community-operators-zfb4t" Dec 06 16:11:19 crc kubenswrapper[4813]: I1206 16:11:19.522047 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cae6355-3939-4aac-9592-3d916f54c6bf-utilities\") pod \"community-operators-zfb4t\" (UID: \"3cae6355-3939-4aac-9592-3d916f54c6bf\") " pod="openshift-marketplace/community-operators-zfb4t" Dec 06 16:11:19 crc kubenswrapper[4813]: I1206 16:11:19.522169 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8v4j\" (UniqueName: \"kubernetes.io/projected/3cae6355-3939-4aac-9592-3d916f54c6bf-kube-api-access-f8v4j\") pod \"community-operators-zfb4t\" (UID: \"3cae6355-3939-4aac-9592-3d916f54c6bf\") " pod="openshift-marketplace/community-operators-zfb4t" Dec 06 16:11:19 crc kubenswrapper[4813]: I1206 16:11:19.624645 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cae6355-3939-4aac-9592-3d916f54c6bf-catalog-content\") pod \"community-operators-zfb4t\" (UID: \"3cae6355-3939-4aac-9592-3d916f54c6bf\") " pod="openshift-marketplace/community-operators-zfb4t" Dec 06 16:11:19 crc kubenswrapper[4813]: I1206 16:11:19.624727 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cae6355-3939-4aac-9592-3d916f54c6bf-utilities\") pod \"community-operators-zfb4t\" (UID: \"3cae6355-3939-4aac-9592-3d916f54c6bf\") " pod="openshift-marketplace/community-operators-zfb4t" Dec 06 16:11:19 crc kubenswrapper[4813]: I1206 16:11:19.624957 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8v4j\" (UniqueName: \"kubernetes.io/projected/3cae6355-3939-4aac-9592-3d916f54c6bf-kube-api-access-f8v4j\") pod \"community-operators-zfb4t\" (UID: \"3cae6355-3939-4aac-9592-3d916f54c6bf\") " pod="openshift-marketplace/community-operators-zfb4t" Dec 06 16:11:19 crc kubenswrapper[4813]: I1206 16:11:19.625403 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cae6355-3939-4aac-9592-3d916f54c6bf-catalog-content\") pod \"community-operators-zfb4t\" (UID: \"3cae6355-3939-4aac-9592-3d916f54c6bf\") " pod="openshift-marketplace/community-operators-zfb4t" Dec 06 16:11:19 crc kubenswrapper[4813]: I1206 16:11:19.626189 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cae6355-3939-4aac-9592-3d916f54c6bf-utilities\") pod \"community-operators-zfb4t\" (UID: \"3cae6355-3939-4aac-9592-3d916f54c6bf\") " pod="openshift-marketplace/community-operators-zfb4t" Dec 06 16:11:19 crc kubenswrapper[4813]: I1206 16:11:19.646470 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8v4j\" (UniqueName: \"kubernetes.io/projected/3cae6355-3939-4aac-9592-3d916f54c6bf-kube-api-access-f8v4j\") pod \"community-operators-zfb4t\" (UID: \"3cae6355-3939-4aac-9592-3d916f54c6bf\") " pod="openshift-marketplace/community-operators-zfb4t" Dec 06 16:11:19 crc kubenswrapper[4813]: I1206 16:11:19.733652 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zfb4t" Dec 06 16:11:20 crc kubenswrapper[4813]: I1206 16:11:20.242829 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zfb4t"] Dec 06 16:11:21 crc kubenswrapper[4813]: I1206 16:11:21.117386 4813 generic.go:334] "Generic (PLEG): container finished" podID="3cae6355-3939-4aac-9592-3d916f54c6bf" containerID="8b47214910b2ee1e105792a2a31f78d815d9cddb6b7eebbc97984127160bb71b" exitCode=0 Dec 06 16:11:21 crc kubenswrapper[4813]: I1206 16:11:21.117441 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfb4t" event={"ID":"3cae6355-3939-4aac-9592-3d916f54c6bf","Type":"ContainerDied","Data":"8b47214910b2ee1e105792a2a31f78d815d9cddb6b7eebbc97984127160bb71b"} Dec 06 16:11:21 crc kubenswrapper[4813]: I1206 16:11:21.117676 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfb4t" event={"ID":"3cae6355-3939-4aac-9592-3d916f54c6bf","Type":"ContainerStarted","Data":"493496999f47d7ca4889318f48df2ee2089deb08432fff82f5a5b027ac72e2de"} Dec 06 16:11:22 crc kubenswrapper[4813]: I1206 16:11:22.128480 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfb4t" event={"ID":"3cae6355-3939-4aac-9592-3d916f54c6bf","Type":"ContainerStarted","Data":"265afc4f3c0c7f2b91414885a1fe239d804bb10aa6d43d7bd933975385713ac5"} Dec 06 16:11:23 crc kubenswrapper[4813]: I1206 16:11:23.147139 4813 generic.go:334] "Generic (PLEG): container finished" podID="3cae6355-3939-4aac-9592-3d916f54c6bf" containerID="265afc4f3c0c7f2b91414885a1fe239d804bb10aa6d43d7bd933975385713ac5" exitCode=0 Dec 06 16:11:23 crc kubenswrapper[4813]: I1206 16:11:23.147314 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfb4t" event={"ID":"3cae6355-3939-4aac-9592-3d916f54c6bf","Type":"ContainerDied","Data":"265afc4f3c0c7f2b91414885a1fe239d804bb10aa6d43d7bd933975385713ac5"} Dec 06 16:11:24 crc kubenswrapper[4813]: I1206 16:11:24.159462 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfb4t" event={"ID":"3cae6355-3939-4aac-9592-3d916f54c6bf","Type":"ContainerStarted","Data":"bc939c506cbe92f840c317d8620a550b71df017d2fa0e87f96a3a50e1f71a05b"} Dec 06 16:11:24 crc kubenswrapper[4813]: I1206 16:11:24.184088 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zfb4t" podStartSLOduration=2.725394445 podStartE2EDuration="5.184063456s" podCreationTimestamp="2025-12-06 16:11:19 +0000 UTC" firstStartedPulling="2025-12-06 16:11:21.120579702 +0000 UTC m=+1521.011459278" lastFinishedPulling="2025-12-06 16:11:23.579248703 +0000 UTC m=+1523.470128289" observedRunningTime="2025-12-06 16:11:24.179153087 +0000 UTC m=+1524.070032673" watchObservedRunningTime="2025-12-06 16:11:24.184063456 +0000 UTC m=+1524.074943042" Dec 06 16:11:28 crc kubenswrapper[4813]: I1206 16:11:28.194872 4813 generic.go:334] "Generic (PLEG): container finished" podID="6306b1cf-bf21-4cd0-a7c8-c83a99863266" containerID="b56853c7861661f6a657ef85062b0040b736154e0fd63d5d7a7e3f0ccd86103c" exitCode=0 Dec 06 16:11:28 crc kubenswrapper[4813]: I1206 16:11:28.194929 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qr892" event={"ID":"6306b1cf-bf21-4cd0-a7c8-c83a99863266","Type":"ContainerDied","Data":"b56853c7861661f6a657ef85062b0040b736154e0fd63d5d7a7e3f0ccd86103c"} Dec 06 16:11:29 crc kubenswrapper[4813]: I1206 16:11:29.695152 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qr892" Dec 06 16:11:29 crc kubenswrapper[4813]: I1206 16:11:29.735515 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zfb4t" Dec 06 16:11:29 crc kubenswrapper[4813]: I1206 16:11:29.736873 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zfb4t" Dec 06 16:11:29 crc kubenswrapper[4813]: I1206 16:11:29.841651 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6306b1cf-bf21-4cd0-a7c8-c83a99863266-inventory\") pod \"6306b1cf-bf21-4cd0-a7c8-c83a99863266\" (UID: \"6306b1cf-bf21-4cd0-a7c8-c83a99863266\") " Dec 06 16:11:29 crc kubenswrapper[4813]: I1206 16:11:29.841946 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6306b1cf-bf21-4cd0-a7c8-c83a99863266-bootstrap-combined-ca-bundle\") pod \"6306b1cf-bf21-4cd0-a7c8-c83a99863266\" (UID: \"6306b1cf-bf21-4cd0-a7c8-c83a99863266\") " Dec 06 16:11:29 crc kubenswrapper[4813]: I1206 16:11:29.842111 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55l5d\" (UniqueName: \"kubernetes.io/projected/6306b1cf-bf21-4cd0-a7c8-c83a99863266-kube-api-access-55l5d\") pod \"6306b1cf-bf21-4cd0-a7c8-c83a99863266\" (UID: \"6306b1cf-bf21-4cd0-a7c8-c83a99863266\") " Dec 06 16:11:29 crc kubenswrapper[4813]: I1206 16:11:29.842198 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6306b1cf-bf21-4cd0-a7c8-c83a99863266-ssh-key\") pod \"6306b1cf-bf21-4cd0-a7c8-c83a99863266\" (UID: \"6306b1cf-bf21-4cd0-a7c8-c83a99863266\") " Dec 06 16:11:29 crc kubenswrapper[4813]: I1206 16:11:29.857470 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6306b1cf-bf21-4cd0-a7c8-c83a99863266-kube-api-access-55l5d" (OuterVolumeSpecName: "kube-api-access-55l5d") pod "6306b1cf-bf21-4cd0-a7c8-c83a99863266" (UID: "6306b1cf-bf21-4cd0-a7c8-c83a99863266"). InnerVolumeSpecName "kube-api-access-55l5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:11:29 crc kubenswrapper[4813]: I1206 16:11:29.870199 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6306b1cf-bf21-4cd0-a7c8-c83a99863266-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "6306b1cf-bf21-4cd0-a7c8-c83a99863266" (UID: "6306b1cf-bf21-4cd0-a7c8-c83a99863266"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:11:29 crc kubenswrapper[4813]: I1206 16:11:29.940585 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6306b1cf-bf21-4cd0-a7c8-c83a99863266-inventory" (OuterVolumeSpecName: "inventory") pod "6306b1cf-bf21-4cd0-a7c8-c83a99863266" (UID: "6306b1cf-bf21-4cd0-a7c8-c83a99863266"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:11:29 crc kubenswrapper[4813]: I1206 16:11:29.943896 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55l5d\" (UniqueName: \"kubernetes.io/projected/6306b1cf-bf21-4cd0-a7c8-c83a99863266-kube-api-access-55l5d\") on node \"crc\" DevicePath \"\"" Dec 06 16:11:29 crc kubenswrapper[4813]: I1206 16:11:29.943919 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6306b1cf-bf21-4cd0-a7c8-c83a99863266-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 16:11:29 crc kubenswrapper[4813]: I1206 16:11:29.943929 4813 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6306b1cf-bf21-4cd0-a7c8-c83a99863266-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:11:29 crc kubenswrapper[4813]: I1206 16:11:29.945946 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6306b1cf-bf21-4cd0-a7c8-c83a99863266-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6306b1cf-bf21-4cd0-a7c8-c83a99863266" (UID: "6306b1cf-bf21-4cd0-a7c8-c83a99863266"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:11:29 crc kubenswrapper[4813]: I1206 16:11:29.948818 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zfb4t" Dec 06 16:11:30 crc kubenswrapper[4813]: I1206 16:11:30.045211 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6306b1cf-bf21-4cd0-a7c8-c83a99863266-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 16:11:30 crc kubenswrapper[4813]: I1206 16:11:30.219950 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qr892" Dec 06 16:11:30 crc kubenswrapper[4813]: I1206 16:11:30.220155 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qr892" event={"ID":"6306b1cf-bf21-4cd0-a7c8-c83a99863266","Type":"ContainerDied","Data":"26b732f26368a1b87e53854bec4306c340399e83779aebe0eb45bbbd1c4b3064"} Dec 06 16:11:30 crc kubenswrapper[4813]: I1206 16:11:30.220300 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26b732f26368a1b87e53854bec4306c340399e83779aebe0eb45bbbd1c4b3064" Dec 06 16:11:30 crc kubenswrapper[4813]: I1206 16:11:30.313386 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zfb4t" Dec 06 16:11:30 crc kubenswrapper[4813]: I1206 16:11:30.332783 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h"] Dec 06 16:11:30 crc kubenswrapper[4813]: E1206 16:11:30.333188 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6306b1cf-bf21-4cd0-a7c8-c83a99863266" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 06 16:11:30 crc kubenswrapper[4813]: I1206 16:11:30.333208 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="6306b1cf-bf21-4cd0-a7c8-c83a99863266" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 06 16:11:30 crc kubenswrapper[4813]: I1206 16:11:30.337745 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="6306b1cf-bf21-4cd0-a7c8-c83a99863266" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 06 16:11:30 crc kubenswrapper[4813]: I1206 16:11:30.338397 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h" Dec 06 16:11:30 crc kubenswrapper[4813]: I1206 16:11:30.344249 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 16:11:30 crc kubenswrapper[4813]: I1206 16:11:30.344486 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 16:11:30 crc kubenswrapper[4813]: I1206 16:11:30.344679 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jvkhf" Dec 06 16:11:30 crc kubenswrapper[4813]: I1206 16:11:30.344564 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 16:11:30 crc kubenswrapper[4813]: I1206 16:11:30.367022 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h"] Dec 06 16:11:30 crc kubenswrapper[4813]: I1206 16:11:30.408940 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zfb4t"] Dec 06 16:11:30 crc kubenswrapper[4813]: I1206 16:11:30.451601 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bswc\" (UniqueName: \"kubernetes.io/projected/c2bb7ef8-9c22-412d-83a9-686036ecb3f8-kube-api-access-2bswc\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h\" (UID: \"c2bb7ef8-9c22-412d-83a9-686036ecb3f8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h" Dec 06 16:11:30 crc kubenswrapper[4813]: I1206 16:11:30.451638 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2bb7ef8-9c22-412d-83a9-686036ecb3f8-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h\" (UID: \"c2bb7ef8-9c22-412d-83a9-686036ecb3f8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h" Dec 06 16:11:30 crc kubenswrapper[4813]: I1206 16:11:30.451724 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2bb7ef8-9c22-412d-83a9-686036ecb3f8-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h\" (UID: \"c2bb7ef8-9c22-412d-83a9-686036ecb3f8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h" Dec 06 16:11:30 crc kubenswrapper[4813]: I1206 16:11:30.553145 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2bb7ef8-9c22-412d-83a9-686036ecb3f8-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h\" (UID: \"c2bb7ef8-9c22-412d-83a9-686036ecb3f8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h" Dec 06 16:11:30 crc kubenswrapper[4813]: I1206 16:11:30.553307 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bswc\" (UniqueName: \"kubernetes.io/projected/c2bb7ef8-9c22-412d-83a9-686036ecb3f8-kube-api-access-2bswc\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h\" (UID: \"c2bb7ef8-9c22-412d-83a9-686036ecb3f8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h" Dec 06 16:11:30 crc kubenswrapper[4813]: I1206 16:11:30.553327 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2bb7ef8-9c22-412d-83a9-686036ecb3f8-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h\" (UID: \"c2bb7ef8-9c22-412d-83a9-686036ecb3f8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h" Dec 06 16:11:30 crc kubenswrapper[4813]: I1206 16:11:30.558008 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2bb7ef8-9c22-412d-83a9-686036ecb3f8-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h\" (UID: \"c2bb7ef8-9c22-412d-83a9-686036ecb3f8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h" Dec 06 16:11:30 crc kubenswrapper[4813]: I1206 16:11:30.558728 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2bb7ef8-9c22-412d-83a9-686036ecb3f8-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h\" (UID: \"c2bb7ef8-9c22-412d-83a9-686036ecb3f8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h" Dec 06 16:11:30 crc kubenswrapper[4813]: I1206 16:11:30.571067 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bswc\" (UniqueName: \"kubernetes.io/projected/c2bb7ef8-9c22-412d-83a9-686036ecb3f8-kube-api-access-2bswc\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h\" (UID: \"c2bb7ef8-9c22-412d-83a9-686036ecb3f8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h" Dec 06 16:11:30 crc kubenswrapper[4813]: I1206 16:11:30.668279 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h" Dec 06 16:11:31 crc kubenswrapper[4813]: I1206 16:11:31.257007 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h"] Dec 06 16:11:32 crc kubenswrapper[4813]: I1206 16:11:32.241742 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h" event={"ID":"c2bb7ef8-9c22-412d-83a9-686036ecb3f8","Type":"ContainerStarted","Data":"6f43d50d4300a398ecafd8123407ed146d1be59545e3d0adfb16ed26bf8f16aa"} Dec 06 16:11:32 crc kubenswrapper[4813]: I1206 16:11:32.242038 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h" event={"ID":"c2bb7ef8-9c22-412d-83a9-686036ecb3f8","Type":"ContainerStarted","Data":"9a523eff8d7384e5635c271a77c4569c6c7448f0d3dcc126e5aea795d57224ee"} Dec 06 16:11:32 crc kubenswrapper[4813]: I1206 16:11:32.242185 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zfb4t" podUID="3cae6355-3939-4aac-9592-3d916f54c6bf" containerName="registry-server" containerID="cri-o://bc939c506cbe92f840c317d8620a550b71df017d2fa0e87f96a3a50e1f71a05b" gracePeriod=2 Dec 06 16:11:32 crc kubenswrapper[4813]: I1206 16:11:32.260735 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h" podStartSLOduration=1.848628803 podStartE2EDuration="2.260717813s" podCreationTimestamp="2025-12-06 16:11:30 +0000 UTC" firstStartedPulling="2025-12-06 16:11:31.265103686 +0000 UTC m=+1531.155983262" lastFinishedPulling="2025-12-06 16:11:31.677192696 +0000 UTC m=+1531.568072272" observedRunningTime="2025-12-06 16:11:32.258584527 +0000 UTC m=+1532.149464103" watchObservedRunningTime="2025-12-06 16:11:32.260717813 +0000 UTC m=+1532.151597389" Dec 06 16:11:32 crc kubenswrapper[4813]: I1206 16:11:32.664859 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zfb4t" Dec 06 16:11:32 crc kubenswrapper[4813]: I1206 16:11:32.791815 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cae6355-3939-4aac-9592-3d916f54c6bf-catalog-content\") pod \"3cae6355-3939-4aac-9592-3d916f54c6bf\" (UID: \"3cae6355-3939-4aac-9592-3d916f54c6bf\") " Dec 06 16:11:32 crc kubenswrapper[4813]: I1206 16:11:32.791924 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8v4j\" (UniqueName: \"kubernetes.io/projected/3cae6355-3939-4aac-9592-3d916f54c6bf-kube-api-access-f8v4j\") pod \"3cae6355-3939-4aac-9592-3d916f54c6bf\" (UID: \"3cae6355-3939-4aac-9592-3d916f54c6bf\") " Dec 06 16:11:32 crc kubenswrapper[4813]: I1206 16:11:32.792061 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cae6355-3939-4aac-9592-3d916f54c6bf-utilities\") pod \"3cae6355-3939-4aac-9592-3d916f54c6bf\" (UID: \"3cae6355-3939-4aac-9592-3d916f54c6bf\") " Dec 06 16:11:32 crc kubenswrapper[4813]: I1206 16:11:32.792917 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cae6355-3939-4aac-9592-3d916f54c6bf-utilities" (OuterVolumeSpecName: "utilities") pod "3cae6355-3939-4aac-9592-3d916f54c6bf" (UID: "3cae6355-3939-4aac-9592-3d916f54c6bf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:11:32 crc kubenswrapper[4813]: I1206 16:11:32.793326 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cae6355-3939-4aac-9592-3d916f54c6bf-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 16:11:32 crc kubenswrapper[4813]: I1206 16:11:32.797429 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cae6355-3939-4aac-9592-3d916f54c6bf-kube-api-access-f8v4j" (OuterVolumeSpecName: "kube-api-access-f8v4j") pod "3cae6355-3939-4aac-9592-3d916f54c6bf" (UID: "3cae6355-3939-4aac-9592-3d916f54c6bf"). InnerVolumeSpecName "kube-api-access-f8v4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:11:32 crc kubenswrapper[4813]: I1206 16:11:32.841536 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cae6355-3939-4aac-9592-3d916f54c6bf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3cae6355-3939-4aac-9592-3d916f54c6bf" (UID: "3cae6355-3939-4aac-9592-3d916f54c6bf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:11:32 crc kubenswrapper[4813]: I1206 16:11:32.894493 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cae6355-3939-4aac-9592-3d916f54c6bf-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 16:11:32 crc kubenswrapper[4813]: I1206 16:11:32.894677 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8v4j\" (UniqueName: \"kubernetes.io/projected/3cae6355-3939-4aac-9592-3d916f54c6bf-kube-api-access-f8v4j\") on node \"crc\" DevicePath \"\"" Dec 06 16:11:33 crc kubenswrapper[4813]: I1206 16:11:33.258240 4813 generic.go:334] "Generic (PLEG): container finished" podID="3cae6355-3939-4aac-9592-3d916f54c6bf" containerID="bc939c506cbe92f840c317d8620a550b71df017d2fa0e87f96a3a50e1f71a05b" exitCode=0 Dec 06 16:11:33 crc kubenswrapper[4813]: I1206 16:11:33.258310 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zfb4t" Dec 06 16:11:33 crc kubenswrapper[4813]: I1206 16:11:33.258375 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfb4t" event={"ID":"3cae6355-3939-4aac-9592-3d916f54c6bf","Type":"ContainerDied","Data":"bc939c506cbe92f840c317d8620a550b71df017d2fa0e87f96a3a50e1f71a05b"} Dec 06 16:11:33 crc kubenswrapper[4813]: I1206 16:11:33.258420 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfb4t" event={"ID":"3cae6355-3939-4aac-9592-3d916f54c6bf","Type":"ContainerDied","Data":"493496999f47d7ca4889318f48df2ee2089deb08432fff82f5a5b027ac72e2de"} Dec 06 16:11:33 crc kubenswrapper[4813]: I1206 16:11:33.258442 4813 scope.go:117] "RemoveContainer" containerID="bc939c506cbe92f840c317d8620a550b71df017d2fa0e87f96a3a50e1f71a05b" Dec 06 16:11:33 crc kubenswrapper[4813]: I1206 16:11:33.316127 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zfb4t"] Dec 06 16:11:33 crc kubenswrapper[4813]: I1206 16:11:33.324173 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zfb4t"] Dec 06 16:11:33 crc kubenswrapper[4813]: I1206 16:11:33.327684 4813 scope.go:117] "RemoveContainer" containerID="265afc4f3c0c7f2b91414885a1fe239d804bb10aa6d43d7bd933975385713ac5" Dec 06 16:11:33 crc kubenswrapper[4813]: I1206 16:11:33.350240 4813 scope.go:117] "RemoveContainer" containerID="8b47214910b2ee1e105792a2a31f78d815d9cddb6b7eebbc97984127160bb71b" Dec 06 16:11:33 crc kubenswrapper[4813]: I1206 16:11:33.388620 4813 scope.go:117] "RemoveContainer" containerID="bc939c506cbe92f840c317d8620a550b71df017d2fa0e87f96a3a50e1f71a05b" Dec 06 16:11:33 crc kubenswrapper[4813]: E1206 16:11:33.390013 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc939c506cbe92f840c317d8620a550b71df017d2fa0e87f96a3a50e1f71a05b\": container with ID starting with bc939c506cbe92f840c317d8620a550b71df017d2fa0e87f96a3a50e1f71a05b not found: ID does not exist" containerID="bc939c506cbe92f840c317d8620a550b71df017d2fa0e87f96a3a50e1f71a05b" Dec 06 16:11:33 crc kubenswrapper[4813]: I1206 16:11:33.390201 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc939c506cbe92f840c317d8620a550b71df017d2fa0e87f96a3a50e1f71a05b"} err="failed to get container status \"bc939c506cbe92f840c317d8620a550b71df017d2fa0e87f96a3a50e1f71a05b\": rpc error: code = NotFound desc = could not find container \"bc939c506cbe92f840c317d8620a550b71df017d2fa0e87f96a3a50e1f71a05b\": container with ID starting with bc939c506cbe92f840c317d8620a550b71df017d2fa0e87f96a3a50e1f71a05b not found: ID does not exist" Dec 06 16:11:33 crc kubenswrapper[4813]: I1206 16:11:33.390238 4813 scope.go:117] "RemoveContainer" containerID="265afc4f3c0c7f2b91414885a1fe239d804bb10aa6d43d7bd933975385713ac5" Dec 06 16:11:33 crc kubenswrapper[4813]: E1206 16:11:33.390768 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"265afc4f3c0c7f2b91414885a1fe239d804bb10aa6d43d7bd933975385713ac5\": container with ID starting with 265afc4f3c0c7f2b91414885a1fe239d804bb10aa6d43d7bd933975385713ac5 not found: ID does not exist" containerID="265afc4f3c0c7f2b91414885a1fe239d804bb10aa6d43d7bd933975385713ac5" Dec 06 16:11:33 crc kubenswrapper[4813]: I1206 16:11:33.390812 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"265afc4f3c0c7f2b91414885a1fe239d804bb10aa6d43d7bd933975385713ac5"} err="failed to get container status \"265afc4f3c0c7f2b91414885a1fe239d804bb10aa6d43d7bd933975385713ac5\": rpc error: code = NotFound desc = could not find container \"265afc4f3c0c7f2b91414885a1fe239d804bb10aa6d43d7bd933975385713ac5\": container with ID starting with 265afc4f3c0c7f2b91414885a1fe239d804bb10aa6d43d7bd933975385713ac5 not found: ID does not exist" Dec 06 16:11:33 crc kubenswrapper[4813]: I1206 16:11:33.390845 4813 scope.go:117] "RemoveContainer" containerID="8b47214910b2ee1e105792a2a31f78d815d9cddb6b7eebbc97984127160bb71b" Dec 06 16:11:33 crc kubenswrapper[4813]: E1206 16:11:33.391111 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b47214910b2ee1e105792a2a31f78d815d9cddb6b7eebbc97984127160bb71b\": container with ID starting with 8b47214910b2ee1e105792a2a31f78d815d9cddb6b7eebbc97984127160bb71b not found: ID does not exist" containerID="8b47214910b2ee1e105792a2a31f78d815d9cddb6b7eebbc97984127160bb71b" Dec 06 16:11:33 crc kubenswrapper[4813]: I1206 16:11:33.391132 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b47214910b2ee1e105792a2a31f78d815d9cddb6b7eebbc97984127160bb71b"} err="failed to get container status \"8b47214910b2ee1e105792a2a31f78d815d9cddb6b7eebbc97984127160bb71b\": rpc error: code = NotFound desc = could not find container \"8b47214910b2ee1e105792a2a31f78d815d9cddb6b7eebbc97984127160bb71b\": container with ID starting with 8b47214910b2ee1e105792a2a31f78d815d9cddb6b7eebbc97984127160bb71b not found: ID does not exist" Dec 06 16:11:34 crc kubenswrapper[4813]: I1206 16:11:34.505443 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cae6355-3939-4aac-9592-3d916f54c6bf" path="/var/lib/kubelet/pods/3cae6355-3939-4aac-9592-3d916f54c6bf/volumes" Dec 06 16:11:49 crc kubenswrapper[4813]: I1206 16:11:49.427515 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:11:49 crc kubenswrapper[4813]: I1206 16:11:49.429796 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:11:57 crc kubenswrapper[4813]: I1206 16:11:57.041429 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-a434-account-create-update-5rxjv"] Dec 06 16:11:57 crc kubenswrapper[4813]: I1206 16:11:57.050455 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-lwz48"] Dec 06 16:11:57 crc kubenswrapper[4813]: I1206 16:11:57.062098 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-a434-account-create-update-5rxjv"] Dec 06 16:11:57 crc kubenswrapper[4813]: I1206 16:11:57.068353 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-lwz48"] Dec 06 16:11:58 crc kubenswrapper[4813]: I1206 16:11:58.049957 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-qcgfq"] Dec 06 16:11:58 crc kubenswrapper[4813]: I1206 16:11:58.064617 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-2d5d-account-create-update-2pzsb"] Dec 06 16:11:58 crc kubenswrapper[4813]: I1206 16:11:58.076684 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-wt55k"] Dec 06 16:11:58 crc kubenswrapper[4813]: I1206 16:11:58.085983 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-qcgfq"] Dec 06 16:11:58 crc kubenswrapper[4813]: I1206 16:11:58.105298 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-c96d-account-create-update-ldmnt"] Dec 06 16:11:58 crc kubenswrapper[4813]: I1206 16:11:58.117198 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-wt55k"] Dec 06 16:11:58 crc kubenswrapper[4813]: I1206 16:11:58.129085 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-2d5d-account-create-update-2pzsb"] Dec 06 16:11:58 crc kubenswrapper[4813]: I1206 16:11:58.141938 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-c96d-account-create-update-ldmnt"] Dec 06 16:11:58 crc kubenswrapper[4813]: I1206 16:11:58.512631 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66c65f92-2424-4d91-9824-9a6c00bf3dbf" path="/var/lib/kubelet/pods/66c65f92-2424-4d91-9824-9a6c00bf3dbf/volumes" Dec 06 16:11:58 crc kubenswrapper[4813]: I1206 16:11:58.515506 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6805af73-82e1-4e15-b867-d02abc99d5ee" path="/var/lib/kubelet/pods/6805af73-82e1-4e15-b867-d02abc99d5ee/volumes" Dec 06 16:11:58 crc kubenswrapper[4813]: I1206 16:11:58.517691 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="854df38a-5bef-4b2c-acdc-4ca7577bc1de" path="/var/lib/kubelet/pods/854df38a-5bef-4b2c-acdc-4ca7577bc1de/volumes" Dec 06 16:11:58 crc kubenswrapper[4813]: I1206 16:11:58.519934 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d882403f-89d4-4bef-b2a0-7f1134e4b735" path="/var/lib/kubelet/pods/d882403f-89d4-4bef-b2a0-7f1134e4b735/volumes" Dec 06 16:11:58 crc kubenswrapper[4813]: I1206 16:11:58.523075 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7dd308a-25f5-4746-b850-d4a2430c6b20" path="/var/lib/kubelet/pods/e7dd308a-25f5-4746-b850-d4a2430c6b20/volumes" Dec 06 16:11:58 crc kubenswrapper[4813]: I1206 16:11:58.524928 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee373829-88aa-427e-aedb-af7410b7b7a3" path="/var/lib/kubelet/pods/ee373829-88aa-427e-aedb-af7410b7b7a3/volumes" Dec 06 16:12:13 crc kubenswrapper[4813]: I1206 16:12:13.223967 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pzrrl"] Dec 06 16:12:13 crc kubenswrapper[4813]: E1206 16:12:13.224762 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cae6355-3939-4aac-9592-3d916f54c6bf" containerName="extract-content" Dec 06 16:12:13 crc kubenswrapper[4813]: I1206 16:12:13.224775 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cae6355-3939-4aac-9592-3d916f54c6bf" containerName="extract-content" Dec 06 16:12:13 crc kubenswrapper[4813]: E1206 16:12:13.224802 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cae6355-3939-4aac-9592-3d916f54c6bf" containerName="registry-server" Dec 06 16:12:13 crc kubenswrapper[4813]: I1206 16:12:13.224809 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cae6355-3939-4aac-9592-3d916f54c6bf" containerName="registry-server" Dec 06 16:12:13 crc kubenswrapper[4813]: E1206 16:12:13.224840 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cae6355-3939-4aac-9592-3d916f54c6bf" containerName="extract-utilities" Dec 06 16:12:13 crc kubenswrapper[4813]: I1206 16:12:13.224847 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cae6355-3939-4aac-9592-3d916f54c6bf" containerName="extract-utilities" Dec 06 16:12:13 crc kubenswrapper[4813]: I1206 16:12:13.225006 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cae6355-3939-4aac-9592-3d916f54c6bf" containerName="registry-server" Dec 06 16:12:13 crc kubenswrapper[4813]: I1206 16:12:13.226214 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pzrrl" Dec 06 16:12:13 crc kubenswrapper[4813]: I1206 16:12:13.260315 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pzrrl"] Dec 06 16:12:13 crc kubenswrapper[4813]: I1206 16:12:13.404272 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk4wq\" (UniqueName: \"kubernetes.io/projected/1de13162-b942-4f71-ae0b-1a040bd40075-kube-api-access-nk4wq\") pod \"redhat-operators-pzrrl\" (UID: \"1de13162-b942-4f71-ae0b-1a040bd40075\") " pod="openshift-marketplace/redhat-operators-pzrrl" Dec 06 16:12:13 crc kubenswrapper[4813]: I1206 16:12:13.404383 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1de13162-b942-4f71-ae0b-1a040bd40075-utilities\") pod \"redhat-operators-pzrrl\" (UID: \"1de13162-b942-4f71-ae0b-1a040bd40075\") " pod="openshift-marketplace/redhat-operators-pzrrl" Dec 06 16:12:13 crc kubenswrapper[4813]: I1206 16:12:13.404443 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1de13162-b942-4f71-ae0b-1a040bd40075-catalog-content\") pod \"redhat-operators-pzrrl\" (UID: \"1de13162-b942-4f71-ae0b-1a040bd40075\") " pod="openshift-marketplace/redhat-operators-pzrrl" Dec 06 16:12:13 crc kubenswrapper[4813]: I1206 16:12:13.506360 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1de13162-b942-4f71-ae0b-1a040bd40075-utilities\") pod \"redhat-operators-pzrrl\" (UID: \"1de13162-b942-4f71-ae0b-1a040bd40075\") " pod="openshift-marketplace/redhat-operators-pzrrl" Dec 06 16:12:13 crc kubenswrapper[4813]: I1206 16:12:13.506461 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1de13162-b942-4f71-ae0b-1a040bd40075-catalog-content\") pod \"redhat-operators-pzrrl\" (UID: \"1de13162-b942-4f71-ae0b-1a040bd40075\") " pod="openshift-marketplace/redhat-operators-pzrrl" Dec 06 16:12:13 crc kubenswrapper[4813]: I1206 16:12:13.506540 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk4wq\" (UniqueName: \"kubernetes.io/projected/1de13162-b942-4f71-ae0b-1a040bd40075-kube-api-access-nk4wq\") pod \"redhat-operators-pzrrl\" (UID: \"1de13162-b942-4f71-ae0b-1a040bd40075\") " pod="openshift-marketplace/redhat-operators-pzrrl" Dec 06 16:12:13 crc kubenswrapper[4813]: I1206 16:12:13.506903 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1de13162-b942-4f71-ae0b-1a040bd40075-utilities\") pod \"redhat-operators-pzrrl\" (UID: \"1de13162-b942-4f71-ae0b-1a040bd40075\") " pod="openshift-marketplace/redhat-operators-pzrrl" Dec 06 16:12:13 crc kubenswrapper[4813]: I1206 16:12:13.506903 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1de13162-b942-4f71-ae0b-1a040bd40075-catalog-content\") pod \"redhat-operators-pzrrl\" (UID: \"1de13162-b942-4f71-ae0b-1a040bd40075\") " pod="openshift-marketplace/redhat-operators-pzrrl" Dec 06 16:12:13 crc kubenswrapper[4813]: I1206 16:12:13.529691 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk4wq\" (UniqueName: \"kubernetes.io/projected/1de13162-b942-4f71-ae0b-1a040bd40075-kube-api-access-nk4wq\") pod \"redhat-operators-pzrrl\" (UID: \"1de13162-b942-4f71-ae0b-1a040bd40075\") " pod="openshift-marketplace/redhat-operators-pzrrl" Dec 06 16:12:13 crc kubenswrapper[4813]: I1206 16:12:13.560247 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pzrrl" Dec 06 16:12:14 crc kubenswrapper[4813]: I1206 16:12:14.087649 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pzrrl"] Dec 06 16:12:14 crc kubenswrapper[4813]: I1206 16:12:14.982044 4813 generic.go:334] "Generic (PLEG): container finished" podID="1de13162-b942-4f71-ae0b-1a040bd40075" containerID="18c644991a76fcf89ab81482429abd09f5515777f443d1b8d7bdf32bef31c7fc" exitCode=0 Dec 06 16:12:14 crc kubenswrapper[4813]: I1206 16:12:14.982151 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pzrrl" event={"ID":"1de13162-b942-4f71-ae0b-1a040bd40075","Type":"ContainerDied","Data":"18c644991a76fcf89ab81482429abd09f5515777f443d1b8d7bdf32bef31c7fc"} Dec 06 16:12:14 crc kubenswrapper[4813]: I1206 16:12:14.982445 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pzrrl" event={"ID":"1de13162-b942-4f71-ae0b-1a040bd40075","Type":"ContainerStarted","Data":"88214f529bdd4f93e20662760d40f8107eccc9434f35b0e098bdbe6968854040"} Dec 06 16:12:15 crc kubenswrapper[4813]: I1206 16:12:15.993209 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pzrrl" event={"ID":"1de13162-b942-4f71-ae0b-1a040bd40075","Type":"ContainerStarted","Data":"d56f4ec18b89376ea4034c47ed0bf162b5b975ac77cb4bbc665aa05872a364b4"} Dec 06 16:12:19 crc kubenswrapper[4813]: I1206 16:12:19.031999 4813 generic.go:334] "Generic (PLEG): container finished" podID="1de13162-b942-4f71-ae0b-1a040bd40075" containerID="d56f4ec18b89376ea4034c47ed0bf162b5b975ac77cb4bbc665aa05872a364b4" exitCode=0 Dec 06 16:12:19 crc kubenswrapper[4813]: I1206 16:12:19.032216 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pzrrl" event={"ID":"1de13162-b942-4f71-ae0b-1a040bd40075","Type":"ContainerDied","Data":"d56f4ec18b89376ea4034c47ed0bf162b5b975ac77cb4bbc665aa05872a364b4"} Dec 06 16:12:19 crc kubenswrapper[4813]: I1206 16:12:19.427960 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:12:19 crc kubenswrapper[4813]: I1206 16:12:19.428034 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:12:19 crc kubenswrapper[4813]: I1206 16:12:19.428084 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" Dec 06 16:12:19 crc kubenswrapper[4813]: I1206 16:12:19.428759 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412"} pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 16:12:19 crc kubenswrapper[4813]: I1206 16:12:19.428825 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" containerID="cri-o://dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412" gracePeriod=600 Dec 06 16:12:19 crc kubenswrapper[4813]: E1206 16:12:19.557994 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:12:20 crc kubenswrapper[4813]: I1206 16:12:20.043069 4813 generic.go:334] "Generic (PLEG): container finished" podID="d88e8bae-c055-4c55-b548-f621ff96de06" containerID="dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412" exitCode=0 Dec 06 16:12:20 crc kubenswrapper[4813]: I1206 16:12:20.043146 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerDied","Data":"dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412"} Dec 06 16:12:20 crc kubenswrapper[4813]: I1206 16:12:20.043216 4813 scope.go:117] "RemoveContainer" containerID="f25008e92574e2c832578c1ff4d4518d3c2eb82b415ff9e3dd99f9504d835fc6" Dec 06 16:12:20 crc kubenswrapper[4813]: I1206 16:12:20.044025 4813 scope.go:117] "RemoveContainer" containerID="dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412" Dec 06 16:12:20 crc kubenswrapper[4813]: E1206 16:12:20.044295 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:12:20 crc kubenswrapper[4813]: I1206 16:12:20.045802 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pzrrl" event={"ID":"1de13162-b942-4f71-ae0b-1a040bd40075","Type":"ContainerStarted","Data":"f4bf7a6c9089aa228538e933c66d40375cb5f504c43f5e2f1ed3d1fc15aa36c3"} Dec 06 16:12:20 crc kubenswrapper[4813]: I1206 16:12:20.091354 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pzrrl" podStartSLOduration=2.660686858 podStartE2EDuration="7.091334374s" podCreationTimestamp="2025-12-06 16:12:13 +0000 UTC" firstStartedPulling="2025-12-06 16:12:14.983489538 +0000 UTC m=+1574.874369104" lastFinishedPulling="2025-12-06 16:12:19.414137034 +0000 UTC m=+1579.305016620" observedRunningTime="2025-12-06 16:12:20.088817319 +0000 UTC m=+1579.979696895" watchObservedRunningTime="2025-12-06 16:12:20.091334374 +0000 UTC m=+1579.982213940" Dec 06 16:12:22 crc kubenswrapper[4813]: I1206 16:12:22.716705 4813 scope.go:117] "RemoveContainer" containerID="af1becfbd82c4ea9d74467e0babd6618407bf7ef7dc9668d2318f06ee35ae8c2" Dec 06 16:12:22 crc kubenswrapper[4813]: I1206 16:12:22.751072 4813 scope.go:117] "RemoveContainer" containerID="97658e2d17b3cfa454f36a100545c61f2a1d47dad3dfa629978409078999cd27" Dec 06 16:12:22 crc kubenswrapper[4813]: I1206 16:12:22.784796 4813 scope.go:117] "RemoveContainer" containerID="bf1a2528259ee03d13821c200c160bd9dc563fcfb453efc86aee5a6e923d1a7c" Dec 06 16:12:22 crc kubenswrapper[4813]: I1206 16:12:22.833590 4813 scope.go:117] "RemoveContainer" containerID="f5b499d0437ecc1d705d99c428fda07eddb536022bca03f7fe21f6896d0106d3" Dec 06 16:12:22 crc kubenswrapper[4813]: I1206 16:12:22.873854 4813 scope.go:117] "RemoveContainer" containerID="ec3f311f852ddce4c0ba7b5c384ed91c57f18949369b7ca524def707555f235e" Dec 06 16:12:22 crc kubenswrapper[4813]: I1206 16:12:22.918686 4813 scope.go:117] "RemoveContainer" containerID="e9b6e23ad4f2a2290ab5e5d6deea14532e1fe49412cd92438393dc10cad29994" Dec 06 16:12:23 crc kubenswrapper[4813]: I1206 16:12:23.561419 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pzrrl" Dec 06 16:12:23 crc kubenswrapper[4813]: I1206 16:12:23.562461 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pzrrl" Dec 06 16:12:24 crc kubenswrapper[4813]: I1206 16:12:24.614007 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pzrrl" podUID="1de13162-b942-4f71-ae0b-1a040bd40075" containerName="registry-server" probeResult="failure" output=< Dec 06 16:12:24 crc kubenswrapper[4813]: timeout: failed to connect service ":50051" within 1s Dec 06 16:12:24 crc kubenswrapper[4813]: > Dec 06 16:12:27 crc kubenswrapper[4813]: I1206 16:12:27.045219 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-4flrx"] Dec 06 16:12:27 crc kubenswrapper[4813]: I1206 16:12:27.056408 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-6nm8x"] Dec 06 16:12:27 crc kubenswrapper[4813]: I1206 16:12:27.067432 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-88qpf"] Dec 06 16:12:27 crc kubenswrapper[4813]: I1206 16:12:27.090593 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-4flrx"] Dec 06 16:12:27 crc kubenswrapper[4813]: I1206 16:12:27.112993 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-88qpf"] Dec 06 16:12:27 crc kubenswrapper[4813]: I1206 16:12:27.124893 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-6nm8x"] Dec 06 16:12:28 crc kubenswrapper[4813]: I1206 16:12:28.506085 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="235e0c8b-b891-4c0c-9676-d0df7a2b6836" path="/var/lib/kubelet/pods/235e0c8b-b891-4c0c-9676-d0df7a2b6836/volumes" Dec 06 16:12:28 crc kubenswrapper[4813]: I1206 16:12:28.508596 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60776264-205f-4541-8d51-3d0d3aec891d" path="/var/lib/kubelet/pods/60776264-205f-4541-8d51-3d0d3aec891d/volumes" Dec 06 16:12:28 crc kubenswrapper[4813]: I1206 16:12:28.510519 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93edeafb-90c8-4ede-b6f4-608b67aadc37" path="/var/lib/kubelet/pods/93edeafb-90c8-4ede-b6f4-608b67aadc37/volumes" Dec 06 16:12:31 crc kubenswrapper[4813]: I1206 16:12:31.041980 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-2f15-account-create-update-2gvw4"] Dec 06 16:12:31 crc kubenswrapper[4813]: I1206 16:12:31.050053 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-tzd8d"] Dec 06 16:12:31 crc kubenswrapper[4813]: I1206 16:12:31.057974 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-24e9-account-create-update-ht2lg"] Dec 06 16:12:31 crc kubenswrapper[4813]: I1206 16:12:31.065081 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-c831-account-create-update-h8lzr"] Dec 06 16:12:31 crc kubenswrapper[4813]: I1206 16:12:31.074697 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-2f15-account-create-update-2gvw4"] Dec 06 16:12:31 crc kubenswrapper[4813]: I1206 16:12:31.082186 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-24e9-account-create-update-ht2lg"] Dec 06 16:12:31 crc kubenswrapper[4813]: I1206 16:12:31.088840 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-tzd8d"] Dec 06 16:12:31 crc kubenswrapper[4813]: I1206 16:12:31.094958 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-c831-account-create-update-h8lzr"] Dec 06 16:12:32 crc kubenswrapper[4813]: I1206 16:12:32.503799 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27b4a8bf-5c37-4936-8ced-6a1d02249dd1" path="/var/lib/kubelet/pods/27b4a8bf-5c37-4936-8ced-6a1d02249dd1/volumes" Dec 06 16:12:32 crc kubenswrapper[4813]: I1206 16:12:32.505182 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3579ce45-95fa-44af-a164-3153a1c873ac" path="/var/lib/kubelet/pods/3579ce45-95fa-44af-a164-3153a1c873ac/volumes" Dec 06 16:12:32 crc kubenswrapper[4813]: I1206 16:12:32.506857 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="612d9b68-b55f-40f5-a45a-c1d7a3b32f2b" path="/var/lib/kubelet/pods/612d9b68-b55f-40f5-a45a-c1d7a3b32f2b/volumes" Dec 06 16:12:32 crc kubenswrapper[4813]: I1206 16:12:32.507997 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9" path="/var/lib/kubelet/pods/9acc6e9c-7b3f-43d4-8b77-e75fca8f85c9/volumes" Dec 06 16:12:33 crc kubenswrapper[4813]: I1206 16:12:33.630988 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pzrrl" Dec 06 16:12:33 crc kubenswrapper[4813]: I1206 16:12:33.697472 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pzrrl" Dec 06 16:12:33 crc kubenswrapper[4813]: I1206 16:12:33.878379 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pzrrl"] Dec 06 16:12:34 crc kubenswrapper[4813]: I1206 16:12:34.487676 4813 scope.go:117] "RemoveContainer" containerID="dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412" Dec 06 16:12:34 crc kubenswrapper[4813]: E1206 16:12:34.488136 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:12:35 crc kubenswrapper[4813]: I1206 16:12:35.044886 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-jz5zm"] Dec 06 16:12:35 crc kubenswrapper[4813]: I1206 16:12:35.056080 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-jz5zm"] Dec 06 16:12:35 crc kubenswrapper[4813]: I1206 16:12:35.204631 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pzrrl" podUID="1de13162-b942-4f71-ae0b-1a040bd40075" containerName="registry-server" containerID="cri-o://f4bf7a6c9089aa228538e933c66d40375cb5f504c43f5e2f1ed3d1fc15aa36c3" gracePeriod=2 Dec 06 16:12:35 crc kubenswrapper[4813]: I1206 16:12:35.621883 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pzrrl" Dec 06 16:12:35 crc kubenswrapper[4813]: I1206 16:12:35.689148 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1de13162-b942-4f71-ae0b-1a040bd40075-catalog-content\") pod \"1de13162-b942-4f71-ae0b-1a040bd40075\" (UID: \"1de13162-b942-4f71-ae0b-1a040bd40075\") " Dec 06 16:12:35 crc kubenswrapper[4813]: I1206 16:12:35.689209 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1de13162-b942-4f71-ae0b-1a040bd40075-utilities\") pod \"1de13162-b942-4f71-ae0b-1a040bd40075\" (UID: \"1de13162-b942-4f71-ae0b-1a040bd40075\") " Dec 06 16:12:35 crc kubenswrapper[4813]: I1206 16:12:35.689373 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nk4wq\" (UniqueName: \"kubernetes.io/projected/1de13162-b942-4f71-ae0b-1a040bd40075-kube-api-access-nk4wq\") pod \"1de13162-b942-4f71-ae0b-1a040bd40075\" (UID: \"1de13162-b942-4f71-ae0b-1a040bd40075\") " Dec 06 16:12:35 crc kubenswrapper[4813]: I1206 16:12:35.690151 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1de13162-b942-4f71-ae0b-1a040bd40075-utilities" (OuterVolumeSpecName: "utilities") pod "1de13162-b942-4f71-ae0b-1a040bd40075" (UID: "1de13162-b942-4f71-ae0b-1a040bd40075"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:12:35 crc kubenswrapper[4813]: I1206 16:12:35.698507 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1de13162-b942-4f71-ae0b-1a040bd40075-kube-api-access-nk4wq" (OuterVolumeSpecName: "kube-api-access-nk4wq") pod "1de13162-b942-4f71-ae0b-1a040bd40075" (UID: "1de13162-b942-4f71-ae0b-1a040bd40075"). InnerVolumeSpecName "kube-api-access-nk4wq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:12:35 crc kubenswrapper[4813]: I1206 16:12:35.791125 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nk4wq\" (UniqueName: \"kubernetes.io/projected/1de13162-b942-4f71-ae0b-1a040bd40075-kube-api-access-nk4wq\") on node \"crc\" DevicePath \"\"" Dec 06 16:12:35 crc kubenswrapper[4813]: I1206 16:12:35.791157 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1de13162-b942-4f71-ae0b-1a040bd40075-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 16:12:35 crc kubenswrapper[4813]: I1206 16:12:35.792846 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1de13162-b942-4f71-ae0b-1a040bd40075-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1de13162-b942-4f71-ae0b-1a040bd40075" (UID: "1de13162-b942-4f71-ae0b-1a040bd40075"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:12:35 crc kubenswrapper[4813]: I1206 16:12:35.893146 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1de13162-b942-4f71-ae0b-1a040bd40075-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 16:12:36 crc kubenswrapper[4813]: I1206 16:12:36.220418 4813 generic.go:334] "Generic (PLEG): container finished" podID="1de13162-b942-4f71-ae0b-1a040bd40075" containerID="f4bf7a6c9089aa228538e933c66d40375cb5f504c43f5e2f1ed3d1fc15aa36c3" exitCode=0 Dec 06 16:12:36 crc kubenswrapper[4813]: I1206 16:12:36.220458 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pzrrl" event={"ID":"1de13162-b942-4f71-ae0b-1a040bd40075","Type":"ContainerDied","Data":"f4bf7a6c9089aa228538e933c66d40375cb5f504c43f5e2f1ed3d1fc15aa36c3"} Dec 06 16:12:36 crc kubenswrapper[4813]: I1206 16:12:36.220484 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pzrrl" event={"ID":"1de13162-b942-4f71-ae0b-1a040bd40075","Type":"ContainerDied","Data":"88214f529bdd4f93e20662760d40f8107eccc9434f35b0e098bdbe6968854040"} Dec 06 16:12:36 crc kubenswrapper[4813]: I1206 16:12:36.220503 4813 scope.go:117] "RemoveContainer" containerID="f4bf7a6c9089aa228538e933c66d40375cb5f504c43f5e2f1ed3d1fc15aa36c3" Dec 06 16:12:36 crc kubenswrapper[4813]: I1206 16:12:36.221669 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pzrrl" Dec 06 16:12:36 crc kubenswrapper[4813]: I1206 16:12:36.280775 4813 scope.go:117] "RemoveContainer" containerID="d56f4ec18b89376ea4034c47ed0bf162b5b975ac77cb4bbc665aa05872a364b4" Dec 06 16:12:36 crc kubenswrapper[4813]: I1206 16:12:36.286426 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pzrrl"] Dec 06 16:12:36 crc kubenswrapper[4813]: I1206 16:12:36.298172 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pzrrl"] Dec 06 16:12:36 crc kubenswrapper[4813]: I1206 16:12:36.313434 4813 scope.go:117] "RemoveContainer" containerID="18c644991a76fcf89ab81482429abd09f5515777f443d1b8d7bdf32bef31c7fc" Dec 06 16:12:36 crc kubenswrapper[4813]: I1206 16:12:36.357253 4813 scope.go:117] "RemoveContainer" containerID="f4bf7a6c9089aa228538e933c66d40375cb5f504c43f5e2f1ed3d1fc15aa36c3" Dec 06 16:12:36 crc kubenswrapper[4813]: E1206 16:12:36.357776 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4bf7a6c9089aa228538e933c66d40375cb5f504c43f5e2f1ed3d1fc15aa36c3\": container with ID starting with f4bf7a6c9089aa228538e933c66d40375cb5f504c43f5e2f1ed3d1fc15aa36c3 not found: ID does not exist" containerID="f4bf7a6c9089aa228538e933c66d40375cb5f504c43f5e2f1ed3d1fc15aa36c3" Dec 06 16:12:36 crc kubenswrapper[4813]: I1206 16:12:36.357829 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4bf7a6c9089aa228538e933c66d40375cb5f504c43f5e2f1ed3d1fc15aa36c3"} err="failed to get container status \"f4bf7a6c9089aa228538e933c66d40375cb5f504c43f5e2f1ed3d1fc15aa36c3\": rpc error: code = NotFound desc = could not find container \"f4bf7a6c9089aa228538e933c66d40375cb5f504c43f5e2f1ed3d1fc15aa36c3\": container with ID starting with f4bf7a6c9089aa228538e933c66d40375cb5f504c43f5e2f1ed3d1fc15aa36c3 not found: ID does not exist" Dec 06 16:12:36 crc kubenswrapper[4813]: I1206 16:12:36.357863 4813 scope.go:117] "RemoveContainer" containerID="d56f4ec18b89376ea4034c47ed0bf162b5b975ac77cb4bbc665aa05872a364b4" Dec 06 16:12:36 crc kubenswrapper[4813]: E1206 16:12:36.359373 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d56f4ec18b89376ea4034c47ed0bf162b5b975ac77cb4bbc665aa05872a364b4\": container with ID starting with d56f4ec18b89376ea4034c47ed0bf162b5b975ac77cb4bbc665aa05872a364b4 not found: ID does not exist" containerID="d56f4ec18b89376ea4034c47ed0bf162b5b975ac77cb4bbc665aa05872a364b4" Dec 06 16:12:36 crc kubenswrapper[4813]: I1206 16:12:36.359445 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d56f4ec18b89376ea4034c47ed0bf162b5b975ac77cb4bbc665aa05872a364b4"} err="failed to get container status \"d56f4ec18b89376ea4034c47ed0bf162b5b975ac77cb4bbc665aa05872a364b4\": rpc error: code = NotFound desc = could not find container \"d56f4ec18b89376ea4034c47ed0bf162b5b975ac77cb4bbc665aa05872a364b4\": container with ID starting with d56f4ec18b89376ea4034c47ed0bf162b5b975ac77cb4bbc665aa05872a364b4 not found: ID does not exist" Dec 06 16:12:36 crc kubenswrapper[4813]: I1206 16:12:36.359473 4813 scope.go:117] "RemoveContainer" containerID="18c644991a76fcf89ab81482429abd09f5515777f443d1b8d7bdf32bef31c7fc" Dec 06 16:12:36 crc kubenswrapper[4813]: E1206 16:12:36.360317 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18c644991a76fcf89ab81482429abd09f5515777f443d1b8d7bdf32bef31c7fc\": container with ID starting with 18c644991a76fcf89ab81482429abd09f5515777f443d1b8d7bdf32bef31c7fc not found: ID does not exist" containerID="18c644991a76fcf89ab81482429abd09f5515777f443d1b8d7bdf32bef31c7fc" Dec 06 16:12:36 crc kubenswrapper[4813]: I1206 16:12:36.360365 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18c644991a76fcf89ab81482429abd09f5515777f443d1b8d7bdf32bef31c7fc"} err="failed to get container status \"18c644991a76fcf89ab81482429abd09f5515777f443d1b8d7bdf32bef31c7fc\": rpc error: code = NotFound desc = could not find container \"18c644991a76fcf89ab81482429abd09f5515777f443d1b8d7bdf32bef31c7fc\": container with ID starting with 18c644991a76fcf89ab81482429abd09f5515777f443d1b8d7bdf32bef31c7fc not found: ID does not exist" Dec 06 16:12:36 crc kubenswrapper[4813]: I1206 16:12:36.500027 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1de13162-b942-4f71-ae0b-1a040bd40075" path="/var/lib/kubelet/pods/1de13162-b942-4f71-ae0b-1a040bd40075/volumes" Dec 06 16:12:36 crc kubenswrapper[4813]: I1206 16:12:36.501219 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cc7028f-2019-4d85-84d9-7e7bb05216ec" path="/var/lib/kubelet/pods/4cc7028f-2019-4d85-84d9-7e7bb05216ec/volumes" Dec 06 16:12:48 crc kubenswrapper[4813]: I1206 16:12:48.487877 4813 scope.go:117] "RemoveContainer" containerID="dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412" Dec 06 16:12:48 crc kubenswrapper[4813]: E1206 16:12:48.490705 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:13:01 crc kubenswrapper[4813]: I1206 16:13:01.486516 4813 scope.go:117] "RemoveContainer" containerID="dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412" Dec 06 16:13:01 crc kubenswrapper[4813]: E1206 16:13:01.487146 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:13:14 crc kubenswrapper[4813]: I1206 16:13:14.065611 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-5zx65"] Dec 06 16:13:14 crc kubenswrapper[4813]: I1206 16:13:14.078468 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-5zx65"] Dec 06 16:13:14 crc kubenswrapper[4813]: I1206 16:13:14.507329 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72532d82-6482-41ac-a461-c792c134dee3" path="/var/lib/kubelet/pods/72532d82-6482-41ac-a461-c792c134dee3/volumes" Dec 06 16:13:16 crc kubenswrapper[4813]: I1206 16:13:16.487407 4813 scope.go:117] "RemoveContainer" containerID="dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412" Dec 06 16:13:16 crc kubenswrapper[4813]: E1206 16:13:16.488233 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:13:20 crc kubenswrapper[4813]: I1206 16:13:20.770211 4813 generic.go:334] "Generic (PLEG): container finished" podID="c2bb7ef8-9c22-412d-83a9-686036ecb3f8" containerID="6f43d50d4300a398ecafd8123407ed146d1be59545e3d0adfb16ed26bf8f16aa" exitCode=0 Dec 06 16:13:20 crc kubenswrapper[4813]: I1206 16:13:20.770279 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h" event={"ID":"c2bb7ef8-9c22-412d-83a9-686036ecb3f8","Type":"ContainerDied","Data":"6f43d50d4300a398ecafd8123407ed146d1be59545e3d0adfb16ed26bf8f16aa"} Dec 06 16:13:22 crc kubenswrapper[4813]: I1206 16:13:22.301132 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h" Dec 06 16:13:22 crc kubenswrapper[4813]: I1206 16:13:22.502531 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2bb7ef8-9c22-412d-83a9-686036ecb3f8-inventory\") pod \"c2bb7ef8-9c22-412d-83a9-686036ecb3f8\" (UID: \"c2bb7ef8-9c22-412d-83a9-686036ecb3f8\") " Dec 06 16:13:22 crc kubenswrapper[4813]: I1206 16:13:22.503013 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2bb7ef8-9c22-412d-83a9-686036ecb3f8-ssh-key\") pod \"c2bb7ef8-9c22-412d-83a9-686036ecb3f8\" (UID: \"c2bb7ef8-9c22-412d-83a9-686036ecb3f8\") " Dec 06 16:13:22 crc kubenswrapper[4813]: I1206 16:13:22.503139 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bswc\" (UniqueName: \"kubernetes.io/projected/c2bb7ef8-9c22-412d-83a9-686036ecb3f8-kube-api-access-2bswc\") pod \"c2bb7ef8-9c22-412d-83a9-686036ecb3f8\" (UID: \"c2bb7ef8-9c22-412d-83a9-686036ecb3f8\") " Dec 06 16:13:22 crc kubenswrapper[4813]: I1206 16:13:22.508714 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2bb7ef8-9c22-412d-83a9-686036ecb3f8-kube-api-access-2bswc" (OuterVolumeSpecName: "kube-api-access-2bswc") pod "c2bb7ef8-9c22-412d-83a9-686036ecb3f8" (UID: "c2bb7ef8-9c22-412d-83a9-686036ecb3f8"). InnerVolumeSpecName "kube-api-access-2bswc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:13:22 crc kubenswrapper[4813]: I1206 16:13:22.538170 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2bb7ef8-9c22-412d-83a9-686036ecb3f8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c2bb7ef8-9c22-412d-83a9-686036ecb3f8" (UID: "c2bb7ef8-9c22-412d-83a9-686036ecb3f8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:13:22 crc kubenswrapper[4813]: I1206 16:13:22.558065 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2bb7ef8-9c22-412d-83a9-686036ecb3f8-inventory" (OuterVolumeSpecName: "inventory") pod "c2bb7ef8-9c22-412d-83a9-686036ecb3f8" (UID: "c2bb7ef8-9c22-412d-83a9-686036ecb3f8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:13:22 crc kubenswrapper[4813]: I1206 16:13:22.605747 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2bb7ef8-9c22-412d-83a9-686036ecb3f8-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 16:13:22 crc kubenswrapper[4813]: I1206 16:13:22.605870 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2bb7ef8-9c22-412d-83a9-686036ecb3f8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 16:13:22 crc kubenswrapper[4813]: I1206 16:13:22.605926 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bswc\" (UniqueName: \"kubernetes.io/projected/c2bb7ef8-9c22-412d-83a9-686036ecb3f8-kube-api-access-2bswc\") on node \"crc\" DevicePath \"\"" Dec 06 16:13:22 crc kubenswrapper[4813]: I1206 16:13:22.787530 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h" event={"ID":"c2bb7ef8-9c22-412d-83a9-686036ecb3f8","Type":"ContainerDied","Data":"9a523eff8d7384e5635c271a77c4569c6c7448f0d3dcc126e5aea795d57224ee"} Dec 06 16:13:22 crc kubenswrapper[4813]: I1206 16:13:22.787563 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a523eff8d7384e5635c271a77c4569c6c7448f0d3dcc126e5aea795d57224ee" Dec 06 16:13:22 crc kubenswrapper[4813]: I1206 16:13:22.787592 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h" Dec 06 16:13:22 crc kubenswrapper[4813]: I1206 16:13:22.900759 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-48kg5"] Dec 06 16:13:22 crc kubenswrapper[4813]: E1206 16:13:22.901071 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1de13162-b942-4f71-ae0b-1a040bd40075" containerName="registry-server" Dec 06 16:13:22 crc kubenswrapper[4813]: I1206 16:13:22.901086 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="1de13162-b942-4f71-ae0b-1a040bd40075" containerName="registry-server" Dec 06 16:13:22 crc kubenswrapper[4813]: E1206 16:13:22.901099 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1de13162-b942-4f71-ae0b-1a040bd40075" containerName="extract-utilities" Dec 06 16:13:22 crc kubenswrapper[4813]: I1206 16:13:22.901105 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="1de13162-b942-4f71-ae0b-1a040bd40075" containerName="extract-utilities" Dec 06 16:13:22 crc kubenswrapper[4813]: E1206 16:13:22.901124 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2bb7ef8-9c22-412d-83a9-686036ecb3f8" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 06 16:13:22 crc kubenswrapper[4813]: I1206 16:13:22.901130 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2bb7ef8-9c22-412d-83a9-686036ecb3f8" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 06 16:13:22 crc kubenswrapper[4813]: E1206 16:13:22.901146 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1de13162-b942-4f71-ae0b-1a040bd40075" containerName="extract-content" Dec 06 16:13:22 crc kubenswrapper[4813]: I1206 16:13:22.901151 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="1de13162-b942-4f71-ae0b-1a040bd40075" containerName="extract-content" Dec 06 16:13:22 crc kubenswrapper[4813]: I1206 16:13:22.901318 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="1de13162-b942-4f71-ae0b-1a040bd40075" containerName="registry-server" Dec 06 16:13:22 crc kubenswrapper[4813]: I1206 16:13:22.901336 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2bb7ef8-9c22-412d-83a9-686036ecb3f8" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 06 16:13:22 crc kubenswrapper[4813]: I1206 16:13:22.901858 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-48kg5" Dec 06 16:13:22 crc kubenswrapper[4813]: I1206 16:13:22.904592 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 16:13:22 crc kubenswrapper[4813]: I1206 16:13:22.904843 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jvkhf" Dec 06 16:13:22 crc kubenswrapper[4813]: I1206 16:13:22.905028 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 16:13:22 crc kubenswrapper[4813]: I1206 16:13:22.909325 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 16:13:22 crc kubenswrapper[4813]: I1206 16:13:22.926456 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-48kg5"] Dec 06 16:13:23 crc kubenswrapper[4813]: I1206 16:13:23.012650 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/11276539-afd1-4649-97eb-867dfcc76819-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-48kg5\" (UID: \"11276539-afd1-4649-97eb-867dfcc76819\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-48kg5" Dec 06 16:13:23 crc kubenswrapper[4813]: I1206 16:13:23.012738 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pppth\" (UniqueName: \"kubernetes.io/projected/11276539-afd1-4649-97eb-867dfcc76819-kube-api-access-pppth\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-48kg5\" (UID: \"11276539-afd1-4649-97eb-867dfcc76819\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-48kg5" Dec 06 16:13:23 crc kubenswrapper[4813]: I1206 16:13:23.012937 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/11276539-afd1-4649-97eb-867dfcc76819-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-48kg5\" (UID: \"11276539-afd1-4649-97eb-867dfcc76819\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-48kg5" Dec 06 16:13:23 crc kubenswrapper[4813]: I1206 16:13:23.043707 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-nrdxf"] Dec 06 16:13:23 crc kubenswrapper[4813]: I1206 16:13:23.050208 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-nrdxf"] Dec 06 16:13:23 crc kubenswrapper[4813]: I1206 16:13:23.088549 4813 scope.go:117] "RemoveContainer" containerID="c421e9528a587793ae5ac6d5f50537230cf1f38bb2b5734e64d67f19d34e4772" Dec 06 16:13:23 crc kubenswrapper[4813]: I1206 16:13:23.114327 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/11276539-afd1-4649-97eb-867dfcc76819-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-48kg5\" (UID: \"11276539-afd1-4649-97eb-867dfcc76819\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-48kg5" Dec 06 16:13:23 crc kubenswrapper[4813]: I1206 16:13:23.114395 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/11276539-afd1-4649-97eb-867dfcc76819-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-48kg5\" (UID: \"11276539-afd1-4649-97eb-867dfcc76819\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-48kg5" Dec 06 16:13:23 crc kubenswrapper[4813]: I1206 16:13:23.114434 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pppth\" (UniqueName: \"kubernetes.io/projected/11276539-afd1-4649-97eb-867dfcc76819-kube-api-access-pppth\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-48kg5\" (UID: \"11276539-afd1-4649-97eb-867dfcc76819\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-48kg5" Dec 06 16:13:23 crc kubenswrapper[4813]: I1206 16:13:23.119183 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/11276539-afd1-4649-97eb-867dfcc76819-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-48kg5\" (UID: \"11276539-afd1-4649-97eb-867dfcc76819\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-48kg5" Dec 06 16:13:23 crc kubenswrapper[4813]: I1206 16:13:23.120408 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/11276539-afd1-4649-97eb-867dfcc76819-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-48kg5\" (UID: \"11276539-afd1-4649-97eb-867dfcc76819\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-48kg5" Dec 06 16:13:23 crc kubenswrapper[4813]: I1206 16:13:23.127440 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pppth\" (UniqueName: \"kubernetes.io/projected/11276539-afd1-4649-97eb-867dfcc76819-kube-api-access-pppth\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-48kg5\" (UID: \"11276539-afd1-4649-97eb-867dfcc76819\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-48kg5" Dec 06 16:13:23 crc kubenswrapper[4813]: I1206 16:13:23.129567 4813 scope.go:117] "RemoveContainer" containerID="7cb1f4efb714e006d797c6e6963e82be21e1b20f8c0a960ee6d4c452adb052c2" Dec 06 16:13:23 crc kubenswrapper[4813]: I1206 16:13:23.149453 4813 scope.go:117] "RemoveContainer" containerID="31c147352c0d0f036388b6a93a22eceb442d0473ce52fc96e9d6e849d15279c2" Dec 06 16:13:23 crc kubenswrapper[4813]: I1206 16:13:23.169211 4813 scope.go:117] "RemoveContainer" containerID="efb2db4848d9528f745baf4875050797406d985b15dee0813fe2f32e0ee85ec3" Dec 06 16:13:23 crc kubenswrapper[4813]: I1206 16:13:23.187621 4813 scope.go:117] "RemoveContainer" containerID="9b4593a84e5a1d8e2f4d1238ec7d3cb54471974e3cf6c166f3e9292376acdf9c" Dec 06 16:13:23 crc kubenswrapper[4813]: I1206 16:13:23.209318 4813 scope.go:117] "RemoveContainer" containerID="612896523b7d109cffa9908510c0d98cc71cce4af36e28f74d2c8d76282ad269" Dec 06 16:13:23 crc kubenswrapper[4813]: I1206 16:13:23.219194 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-48kg5" Dec 06 16:13:23 crc kubenswrapper[4813]: I1206 16:13:23.406006 4813 scope.go:117] "RemoveContainer" containerID="154a8dd12aee5a84bc8121d41fa15b0771b23a193b0423a6ed68c69d2503a638" Dec 06 16:13:23 crc kubenswrapper[4813]: I1206 16:13:23.444014 4813 scope.go:117] "RemoveContainer" containerID="173f5192c5dbe3f5cea4ce51a3efa20d5cdf784654ac5b83dc8ccaeaa6f592cb" Dec 06 16:13:23 crc kubenswrapper[4813]: I1206 16:13:23.496205 4813 scope.go:117] "RemoveContainer" containerID="1ef292affd710b39999c751c2f5a2ca32230e936cfa573d521c6d139e9f27c7c" Dec 06 16:13:23 crc kubenswrapper[4813]: I1206 16:13:23.845806 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-48kg5"] Dec 06 16:13:24 crc kubenswrapper[4813]: I1206 16:13:24.033602 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-qk8sl"] Dec 06 16:13:24 crc kubenswrapper[4813]: I1206 16:13:24.045048 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-qk8sl"] Dec 06 16:13:24 crc kubenswrapper[4813]: I1206 16:13:24.496508 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0179e5b2-8549-4810-83ab-0acbff2a5c6c" path="/var/lib/kubelet/pods/0179e5b2-8549-4810-83ab-0acbff2a5c6c/volumes" Dec 06 16:13:24 crc kubenswrapper[4813]: I1206 16:13:24.497742 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c1f8568-2368-445f-a2cb-01ab3a6d1564" path="/var/lib/kubelet/pods/5c1f8568-2368-445f-a2cb-01ab3a6d1564/volumes" Dec 06 16:13:24 crc kubenswrapper[4813]: I1206 16:13:24.812046 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-48kg5" event={"ID":"11276539-afd1-4649-97eb-867dfcc76819","Type":"ContainerStarted","Data":"1930ee940ad689da41479da5e60e9fc0901045822ae08fc7f973e8af65e77ade"} Dec 06 16:13:24 crc kubenswrapper[4813]: I1206 16:13:24.812095 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-48kg5" event={"ID":"11276539-afd1-4649-97eb-867dfcc76819","Type":"ContainerStarted","Data":"1bd4e7840f9730353381eb7383ee2ef35a2bdc2b6da49c4b3a5f05ce2db59e71"} Dec 06 16:13:24 crc kubenswrapper[4813]: I1206 16:13:24.839101 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-48kg5" podStartSLOduration=2.447115696 podStartE2EDuration="2.839074159s" podCreationTimestamp="2025-12-06 16:13:22 +0000 UTC" firstStartedPulling="2025-12-06 16:13:23.853957617 +0000 UTC m=+1643.744837203" lastFinishedPulling="2025-12-06 16:13:24.24591606 +0000 UTC m=+1644.136795666" observedRunningTime="2025-12-06 16:13:24.830702653 +0000 UTC m=+1644.721582269" watchObservedRunningTime="2025-12-06 16:13:24.839074159 +0000 UTC m=+1644.729953775" Dec 06 16:13:27 crc kubenswrapper[4813]: I1206 16:13:27.486381 4813 scope.go:117] "RemoveContainer" containerID="dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412" Dec 06 16:13:27 crc kubenswrapper[4813]: E1206 16:13:27.486689 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:13:32 crc kubenswrapper[4813]: I1206 16:13:32.060509 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-6qjfq"] Dec 06 16:13:32 crc kubenswrapper[4813]: I1206 16:13:32.067667 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-6qjfq"] Dec 06 16:13:32 crc kubenswrapper[4813]: I1206 16:13:32.499753 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67cf9ebe-1a58-4e50-9c10-f6f7933e781f" path="/var/lib/kubelet/pods/67cf9ebe-1a58-4e50-9c10-f6f7933e781f/volumes" Dec 06 16:13:38 crc kubenswrapper[4813]: I1206 16:13:38.487889 4813 scope.go:117] "RemoveContainer" containerID="dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412" Dec 06 16:13:38 crc kubenswrapper[4813]: E1206 16:13:38.489045 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:13:48 crc kubenswrapper[4813]: I1206 16:13:48.057232 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-lvw7n"] Dec 06 16:13:48 crc kubenswrapper[4813]: I1206 16:13:48.071072 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-lvw7n"] Dec 06 16:13:48 crc kubenswrapper[4813]: I1206 16:13:48.507003 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cf4a1cf-ef96-4e50-946c-cd0c5eccff32" path="/var/lib/kubelet/pods/8cf4a1cf-ef96-4e50-946c-cd0c5eccff32/volumes" Dec 06 16:13:50 crc kubenswrapper[4813]: I1206 16:13:50.505982 4813 scope.go:117] "RemoveContainer" containerID="dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412" Dec 06 16:13:50 crc kubenswrapper[4813]: E1206 16:13:50.506395 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:14:01 crc kubenswrapper[4813]: I1206 16:14:01.486575 4813 scope.go:117] "RemoveContainer" containerID="dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412" Dec 06 16:14:01 crc kubenswrapper[4813]: E1206 16:14:01.487241 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:14:12 crc kubenswrapper[4813]: I1206 16:14:12.486747 4813 scope.go:117] "RemoveContainer" containerID="dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412" Dec 06 16:14:12 crc kubenswrapper[4813]: E1206 16:14:12.487361 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:14:23 crc kubenswrapper[4813]: I1206 16:14:23.717750 4813 scope.go:117] "RemoveContainer" containerID="31c053cc89a499a51c9bf185c2591dc385e1869e2770b4367925745f679c2804" Dec 06 16:14:23 crc kubenswrapper[4813]: I1206 16:14:23.775132 4813 scope.go:117] "RemoveContainer" containerID="0a3cd73bf8fefc94389973058f01beb8ba5cf10cdc7fb883c3651d2a34f9ed30" Dec 06 16:14:23 crc kubenswrapper[4813]: I1206 16:14:23.836086 4813 scope.go:117] "RemoveContainer" containerID="6033edd6c54603f70163c37784357d9aa1b8fa35bfeb8579b6476ddaaabc69cf" Dec 06 16:14:23 crc kubenswrapper[4813]: I1206 16:14:23.865191 4813 scope.go:117] "RemoveContainer" containerID="e8a3382caedf6fb750d7213c7c4ffe6af9b5c1791f066d46a29aff583c43e72a" Dec 06 16:14:24 crc kubenswrapper[4813]: I1206 16:14:24.487577 4813 scope.go:117] "RemoveContainer" containerID="dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412" Dec 06 16:14:24 crc kubenswrapper[4813]: E1206 16:14:24.488064 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:14:31 crc kubenswrapper[4813]: I1206 16:14:31.044964 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-3fbf-account-create-update-ksdbr"] Dec 06 16:14:31 crc kubenswrapper[4813]: I1206 16:14:31.061211 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-t86b9"] Dec 06 16:14:31 crc kubenswrapper[4813]: I1206 16:14:31.072378 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-3fbf-account-create-update-ksdbr"] Dec 06 16:14:31 crc kubenswrapper[4813]: I1206 16:14:31.082545 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-t86b9"] Dec 06 16:14:32 crc kubenswrapper[4813]: I1206 16:14:32.034472 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-r2fvl"] Dec 06 16:14:32 crc kubenswrapper[4813]: I1206 16:14:32.048346 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-mxrr6"] Dec 06 16:14:32 crc kubenswrapper[4813]: I1206 16:14:32.057365 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-56b5-account-create-update-98zzg"] Dec 06 16:14:32 crc kubenswrapper[4813]: I1206 16:14:32.064381 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db9a-account-create-update-kb927"] Dec 06 16:14:32 crc kubenswrapper[4813]: I1206 16:14:32.072723 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-r2fvl"] Dec 06 16:14:32 crc kubenswrapper[4813]: I1206 16:14:32.080249 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db9a-account-create-update-kb927"] Dec 06 16:14:32 crc kubenswrapper[4813]: I1206 16:14:32.087358 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-mxrr6"] Dec 06 16:14:32 crc kubenswrapper[4813]: I1206 16:14:32.094109 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-56b5-account-create-update-98zzg"] Dec 06 16:14:32 crc kubenswrapper[4813]: I1206 16:14:32.499360 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47a481f6-b873-4920-a273-6e65f8d6503a" path="/var/lib/kubelet/pods/47a481f6-b873-4920-a273-6e65f8d6503a/volumes" Dec 06 16:14:32 crc kubenswrapper[4813]: I1206 16:14:32.500893 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="532ac93c-0c2c-4bca-b9f2-e4f066913dbf" path="/var/lib/kubelet/pods/532ac93c-0c2c-4bca-b9f2-e4f066913dbf/volumes" Dec 06 16:14:32 crc kubenswrapper[4813]: I1206 16:14:32.502107 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d6f9192-b64d-4390-a4bd-4ff0a33bc31b" path="/var/lib/kubelet/pods/5d6f9192-b64d-4390-a4bd-4ff0a33bc31b/volumes" Dec 06 16:14:32 crc kubenswrapper[4813]: I1206 16:14:32.504855 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fbe6119-47a9-4188-b8c0-a9e87d8ea09f" path="/var/lib/kubelet/pods/7fbe6119-47a9-4188-b8c0-a9e87d8ea09f/volumes" Dec 06 16:14:32 crc kubenswrapper[4813]: I1206 16:14:32.506522 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a06cadf5-56f1-439f-8deb-2aaedba66c8a" path="/var/lib/kubelet/pods/a06cadf5-56f1-439f-8deb-2aaedba66c8a/volumes" Dec 06 16:14:32 crc kubenswrapper[4813]: I1206 16:14:32.507480 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc35c201-8f5a-418b-a615-c04dff409d8d" path="/var/lib/kubelet/pods/dc35c201-8f5a-418b-a615-c04dff409d8d/volumes" Dec 06 16:14:38 crc kubenswrapper[4813]: I1206 16:14:38.487756 4813 scope.go:117] "RemoveContainer" containerID="dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412" Dec 06 16:14:38 crc kubenswrapper[4813]: E1206 16:14:38.488808 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:14:42 crc kubenswrapper[4813]: I1206 16:14:42.781286 4813 generic.go:334] "Generic (PLEG): container finished" podID="11276539-afd1-4649-97eb-867dfcc76819" containerID="1930ee940ad689da41479da5e60e9fc0901045822ae08fc7f973e8af65e77ade" exitCode=0 Dec 06 16:14:42 crc kubenswrapper[4813]: I1206 16:14:42.781690 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-48kg5" event={"ID":"11276539-afd1-4649-97eb-867dfcc76819","Type":"ContainerDied","Data":"1930ee940ad689da41479da5e60e9fc0901045822ae08fc7f973e8af65e77ade"} Dec 06 16:14:44 crc kubenswrapper[4813]: I1206 16:14:44.179013 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-48kg5" Dec 06 16:14:44 crc kubenswrapper[4813]: I1206 16:14:44.246806 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pppth\" (UniqueName: \"kubernetes.io/projected/11276539-afd1-4649-97eb-867dfcc76819-kube-api-access-pppth\") pod \"11276539-afd1-4649-97eb-867dfcc76819\" (UID: \"11276539-afd1-4649-97eb-867dfcc76819\") " Dec 06 16:14:44 crc kubenswrapper[4813]: I1206 16:14:44.246972 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/11276539-afd1-4649-97eb-867dfcc76819-ssh-key\") pod \"11276539-afd1-4649-97eb-867dfcc76819\" (UID: \"11276539-afd1-4649-97eb-867dfcc76819\") " Dec 06 16:14:44 crc kubenswrapper[4813]: I1206 16:14:44.247117 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/11276539-afd1-4649-97eb-867dfcc76819-inventory\") pod \"11276539-afd1-4649-97eb-867dfcc76819\" (UID: \"11276539-afd1-4649-97eb-867dfcc76819\") " Dec 06 16:14:44 crc kubenswrapper[4813]: I1206 16:14:44.255399 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11276539-afd1-4649-97eb-867dfcc76819-kube-api-access-pppth" (OuterVolumeSpecName: "kube-api-access-pppth") pod "11276539-afd1-4649-97eb-867dfcc76819" (UID: "11276539-afd1-4649-97eb-867dfcc76819"). InnerVolumeSpecName "kube-api-access-pppth". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:14:44 crc kubenswrapper[4813]: I1206 16:14:44.279375 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11276539-afd1-4649-97eb-867dfcc76819-inventory" (OuterVolumeSpecName: "inventory") pod "11276539-afd1-4649-97eb-867dfcc76819" (UID: "11276539-afd1-4649-97eb-867dfcc76819"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:14:44 crc kubenswrapper[4813]: I1206 16:14:44.285009 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11276539-afd1-4649-97eb-867dfcc76819-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "11276539-afd1-4649-97eb-867dfcc76819" (UID: "11276539-afd1-4649-97eb-867dfcc76819"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:14:44 crc kubenswrapper[4813]: I1206 16:14:44.349571 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/11276539-afd1-4649-97eb-867dfcc76819-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 16:14:44 crc kubenswrapper[4813]: I1206 16:14:44.349802 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pppth\" (UniqueName: \"kubernetes.io/projected/11276539-afd1-4649-97eb-867dfcc76819-kube-api-access-pppth\") on node \"crc\" DevicePath \"\"" Dec 06 16:14:44 crc kubenswrapper[4813]: I1206 16:14:44.349890 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/11276539-afd1-4649-97eb-867dfcc76819-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 16:14:44 crc kubenswrapper[4813]: I1206 16:14:44.812693 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-48kg5" event={"ID":"11276539-afd1-4649-97eb-867dfcc76819","Type":"ContainerDied","Data":"1bd4e7840f9730353381eb7383ee2ef35a2bdc2b6da49c4b3a5f05ce2db59e71"} Dec 06 16:14:44 crc kubenswrapper[4813]: I1206 16:14:44.812727 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1bd4e7840f9730353381eb7383ee2ef35a2bdc2b6da49c4b3a5f05ce2db59e71" Dec 06 16:14:44 crc kubenswrapper[4813]: I1206 16:14:44.812832 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-48kg5" Dec 06 16:14:44 crc kubenswrapper[4813]: I1206 16:14:44.929964 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5"] Dec 06 16:14:44 crc kubenswrapper[4813]: E1206 16:14:44.930566 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11276539-afd1-4649-97eb-867dfcc76819" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 06 16:14:44 crc kubenswrapper[4813]: I1206 16:14:44.930591 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="11276539-afd1-4649-97eb-867dfcc76819" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 06 16:14:44 crc kubenswrapper[4813]: I1206 16:14:44.930902 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="11276539-afd1-4649-97eb-867dfcc76819" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 06 16:14:44 crc kubenswrapper[4813]: I1206 16:14:44.931662 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5" Dec 06 16:14:44 crc kubenswrapper[4813]: I1206 16:14:44.934936 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 16:14:44 crc kubenswrapper[4813]: I1206 16:14:44.935377 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 16:14:44 crc kubenswrapper[4813]: I1206 16:14:44.935698 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jvkhf" Dec 06 16:14:44 crc kubenswrapper[4813]: I1206 16:14:44.940121 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5"] Dec 06 16:14:44 crc kubenswrapper[4813]: I1206 16:14:44.940678 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 16:14:44 crc kubenswrapper[4813]: I1206 16:14:44.970251 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07e3bef8-6975-428b-ac2e-265d2db3d70f-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5\" (UID: \"07e3bef8-6975-428b-ac2e-265d2db3d70f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5" Dec 06 16:14:44 crc kubenswrapper[4813]: I1206 16:14:44.970823 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vv2xs\" (UniqueName: \"kubernetes.io/projected/07e3bef8-6975-428b-ac2e-265d2db3d70f-kube-api-access-vv2xs\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5\" (UID: \"07e3bef8-6975-428b-ac2e-265d2db3d70f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5" Dec 06 16:14:44 crc kubenswrapper[4813]: I1206 16:14:44.970970 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07e3bef8-6975-428b-ac2e-265d2db3d70f-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5\" (UID: \"07e3bef8-6975-428b-ac2e-265d2db3d70f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5" Dec 06 16:14:45 crc kubenswrapper[4813]: I1206 16:14:45.071709 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07e3bef8-6975-428b-ac2e-265d2db3d70f-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5\" (UID: \"07e3bef8-6975-428b-ac2e-265d2db3d70f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5" Dec 06 16:14:45 crc kubenswrapper[4813]: I1206 16:14:45.072077 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07e3bef8-6975-428b-ac2e-265d2db3d70f-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5\" (UID: \"07e3bef8-6975-428b-ac2e-265d2db3d70f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5" Dec 06 16:14:45 crc kubenswrapper[4813]: I1206 16:14:45.072196 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vv2xs\" (UniqueName: \"kubernetes.io/projected/07e3bef8-6975-428b-ac2e-265d2db3d70f-kube-api-access-vv2xs\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5\" (UID: \"07e3bef8-6975-428b-ac2e-265d2db3d70f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5" Dec 06 16:14:45 crc kubenswrapper[4813]: I1206 16:14:45.078546 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07e3bef8-6975-428b-ac2e-265d2db3d70f-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5\" (UID: \"07e3bef8-6975-428b-ac2e-265d2db3d70f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5" Dec 06 16:14:45 crc kubenswrapper[4813]: I1206 16:14:45.086677 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07e3bef8-6975-428b-ac2e-265d2db3d70f-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5\" (UID: \"07e3bef8-6975-428b-ac2e-265d2db3d70f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5" Dec 06 16:14:45 crc kubenswrapper[4813]: I1206 16:14:45.119506 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vv2xs\" (UniqueName: \"kubernetes.io/projected/07e3bef8-6975-428b-ac2e-265d2db3d70f-kube-api-access-vv2xs\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5\" (UID: \"07e3bef8-6975-428b-ac2e-265d2db3d70f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5" Dec 06 16:14:45 crc kubenswrapper[4813]: I1206 16:14:45.265812 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5" Dec 06 16:14:45 crc kubenswrapper[4813]: I1206 16:14:45.813533 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5"] Dec 06 16:14:45 crc kubenswrapper[4813]: I1206 16:14:45.822030 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5" event={"ID":"07e3bef8-6975-428b-ac2e-265d2db3d70f","Type":"ContainerStarted","Data":"9d111ca3de6e4dd2e56129056d0dbb8e6b5e4e6552f2422f2c96f9d44d6cd472"} Dec 06 16:14:46 crc kubenswrapper[4813]: I1206 16:14:46.833033 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5" event={"ID":"07e3bef8-6975-428b-ac2e-265d2db3d70f","Type":"ContainerStarted","Data":"1eaa455f812d1034708b588a36e8caf192071e9761a24d97ee5bc7491d4daed3"} Dec 06 16:14:46 crc kubenswrapper[4813]: I1206 16:14:46.858358 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5" podStartSLOduration=2.361512389 podStartE2EDuration="2.858332056s" podCreationTimestamp="2025-12-06 16:14:44 +0000 UTC" firstStartedPulling="2025-12-06 16:14:45.809468533 +0000 UTC m=+1725.700348139" lastFinishedPulling="2025-12-06 16:14:46.30628822 +0000 UTC m=+1726.197167806" observedRunningTime="2025-12-06 16:14:46.853686581 +0000 UTC m=+1726.744566157" watchObservedRunningTime="2025-12-06 16:14:46.858332056 +0000 UTC m=+1726.749211642" Dec 06 16:14:49 crc kubenswrapper[4813]: I1206 16:14:49.487126 4813 scope.go:117] "RemoveContainer" containerID="dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412" Dec 06 16:14:49 crc kubenswrapper[4813]: E1206 16:14:49.488585 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:14:51 crc kubenswrapper[4813]: I1206 16:14:51.893788 4813 generic.go:334] "Generic (PLEG): container finished" podID="07e3bef8-6975-428b-ac2e-265d2db3d70f" containerID="1eaa455f812d1034708b588a36e8caf192071e9761a24d97ee5bc7491d4daed3" exitCode=0 Dec 06 16:14:51 crc kubenswrapper[4813]: I1206 16:14:51.893931 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5" event={"ID":"07e3bef8-6975-428b-ac2e-265d2db3d70f","Type":"ContainerDied","Data":"1eaa455f812d1034708b588a36e8caf192071e9761a24d97ee5bc7491d4daed3"} Dec 06 16:14:53 crc kubenswrapper[4813]: I1206 16:14:53.427615 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5" Dec 06 16:14:53 crc kubenswrapper[4813]: I1206 16:14:53.566077 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07e3bef8-6975-428b-ac2e-265d2db3d70f-inventory\") pod \"07e3bef8-6975-428b-ac2e-265d2db3d70f\" (UID: \"07e3bef8-6975-428b-ac2e-265d2db3d70f\") " Dec 06 16:14:53 crc kubenswrapper[4813]: I1206 16:14:53.566283 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vv2xs\" (UniqueName: \"kubernetes.io/projected/07e3bef8-6975-428b-ac2e-265d2db3d70f-kube-api-access-vv2xs\") pod \"07e3bef8-6975-428b-ac2e-265d2db3d70f\" (UID: \"07e3bef8-6975-428b-ac2e-265d2db3d70f\") " Dec 06 16:14:53 crc kubenswrapper[4813]: I1206 16:14:53.566308 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07e3bef8-6975-428b-ac2e-265d2db3d70f-ssh-key\") pod \"07e3bef8-6975-428b-ac2e-265d2db3d70f\" (UID: \"07e3bef8-6975-428b-ac2e-265d2db3d70f\") " Dec 06 16:14:53 crc kubenswrapper[4813]: I1206 16:14:53.583493 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07e3bef8-6975-428b-ac2e-265d2db3d70f-kube-api-access-vv2xs" (OuterVolumeSpecName: "kube-api-access-vv2xs") pod "07e3bef8-6975-428b-ac2e-265d2db3d70f" (UID: "07e3bef8-6975-428b-ac2e-265d2db3d70f"). InnerVolumeSpecName "kube-api-access-vv2xs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:14:53 crc kubenswrapper[4813]: I1206 16:14:53.598014 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07e3bef8-6975-428b-ac2e-265d2db3d70f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "07e3bef8-6975-428b-ac2e-265d2db3d70f" (UID: "07e3bef8-6975-428b-ac2e-265d2db3d70f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:14:53 crc kubenswrapper[4813]: I1206 16:14:53.598473 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07e3bef8-6975-428b-ac2e-265d2db3d70f-inventory" (OuterVolumeSpecName: "inventory") pod "07e3bef8-6975-428b-ac2e-265d2db3d70f" (UID: "07e3bef8-6975-428b-ac2e-265d2db3d70f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:14:53 crc kubenswrapper[4813]: I1206 16:14:53.668451 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07e3bef8-6975-428b-ac2e-265d2db3d70f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 16:14:53 crc kubenswrapper[4813]: I1206 16:14:53.668688 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07e3bef8-6975-428b-ac2e-265d2db3d70f-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 16:14:53 crc kubenswrapper[4813]: I1206 16:14:53.668702 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vv2xs\" (UniqueName: \"kubernetes.io/projected/07e3bef8-6975-428b-ac2e-265d2db3d70f-kube-api-access-vv2xs\") on node \"crc\" DevicePath \"\"" Dec 06 16:14:53 crc kubenswrapper[4813]: I1206 16:14:53.921521 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5" event={"ID":"07e3bef8-6975-428b-ac2e-265d2db3d70f","Type":"ContainerDied","Data":"9d111ca3de6e4dd2e56129056d0dbb8e6b5e4e6552f2422f2c96f9d44d6cd472"} Dec 06 16:14:53 crc kubenswrapper[4813]: I1206 16:14:53.921578 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d111ca3de6e4dd2e56129056d0dbb8e6b5e4e6552f2422f2c96f9d44d6cd472" Dec 06 16:14:53 crc kubenswrapper[4813]: I1206 16:14:53.921626 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5" Dec 06 16:14:54 crc kubenswrapper[4813]: I1206 16:14:54.054464 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-nbw6k"] Dec 06 16:14:54 crc kubenswrapper[4813]: E1206 16:14:54.056483 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07e3bef8-6975-428b-ac2e-265d2db3d70f" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 06 16:14:54 crc kubenswrapper[4813]: I1206 16:14:54.056512 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="07e3bef8-6975-428b-ac2e-265d2db3d70f" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 06 16:14:54 crc kubenswrapper[4813]: I1206 16:14:54.056798 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="07e3bef8-6975-428b-ac2e-265d2db3d70f" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 06 16:14:54 crc kubenswrapper[4813]: I1206 16:14:54.057840 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nbw6k" Dec 06 16:14:54 crc kubenswrapper[4813]: I1206 16:14:54.065137 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 16:14:54 crc kubenswrapper[4813]: I1206 16:14:54.069072 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 16:14:54 crc kubenswrapper[4813]: I1206 16:14:54.069530 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jvkhf" Dec 06 16:14:54 crc kubenswrapper[4813]: I1206 16:14:54.075080 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 16:14:54 crc kubenswrapper[4813]: I1206 16:14:54.079448 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-nbw6k"] Dec 06 16:14:54 crc kubenswrapper[4813]: I1206 16:14:54.178443 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c8332abd-e13b-4587-831b-1330e18de573-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nbw6k\" (UID: \"c8332abd-e13b-4587-831b-1330e18de573\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nbw6k" Dec 06 16:14:54 crc kubenswrapper[4813]: I1206 16:14:54.178511 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c8332abd-e13b-4587-831b-1330e18de573-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nbw6k\" (UID: \"c8332abd-e13b-4587-831b-1330e18de573\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nbw6k" Dec 06 16:14:54 crc kubenswrapper[4813]: I1206 16:14:54.178637 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wttqb\" (UniqueName: \"kubernetes.io/projected/c8332abd-e13b-4587-831b-1330e18de573-kube-api-access-wttqb\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nbw6k\" (UID: \"c8332abd-e13b-4587-831b-1330e18de573\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nbw6k" Dec 06 16:14:54 crc kubenswrapper[4813]: I1206 16:14:54.280849 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wttqb\" (UniqueName: \"kubernetes.io/projected/c8332abd-e13b-4587-831b-1330e18de573-kube-api-access-wttqb\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nbw6k\" (UID: \"c8332abd-e13b-4587-831b-1330e18de573\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nbw6k" Dec 06 16:14:54 crc kubenswrapper[4813]: I1206 16:14:54.280929 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c8332abd-e13b-4587-831b-1330e18de573-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nbw6k\" (UID: \"c8332abd-e13b-4587-831b-1330e18de573\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nbw6k" Dec 06 16:14:54 crc kubenswrapper[4813]: I1206 16:14:54.280974 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c8332abd-e13b-4587-831b-1330e18de573-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nbw6k\" (UID: \"c8332abd-e13b-4587-831b-1330e18de573\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nbw6k" Dec 06 16:14:54 crc kubenswrapper[4813]: I1206 16:14:54.286921 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c8332abd-e13b-4587-831b-1330e18de573-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nbw6k\" (UID: \"c8332abd-e13b-4587-831b-1330e18de573\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nbw6k" Dec 06 16:14:54 crc kubenswrapper[4813]: I1206 16:14:54.290111 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c8332abd-e13b-4587-831b-1330e18de573-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nbw6k\" (UID: \"c8332abd-e13b-4587-831b-1330e18de573\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nbw6k" Dec 06 16:14:54 crc kubenswrapper[4813]: I1206 16:14:54.309953 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wttqb\" (UniqueName: \"kubernetes.io/projected/c8332abd-e13b-4587-831b-1330e18de573-kube-api-access-wttqb\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nbw6k\" (UID: \"c8332abd-e13b-4587-831b-1330e18de573\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nbw6k" Dec 06 16:14:54 crc kubenswrapper[4813]: I1206 16:14:54.427310 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nbw6k" Dec 06 16:14:55 crc kubenswrapper[4813]: I1206 16:14:55.331828 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-nbw6k"] Dec 06 16:14:55 crc kubenswrapper[4813]: I1206 16:14:55.937370 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nbw6k" event={"ID":"c8332abd-e13b-4587-831b-1330e18de573","Type":"ContainerStarted","Data":"39e00e030492b64bb72bac3a48f46d57f748455521b327a65577cbfbe0e26346"} Dec 06 16:14:56 crc kubenswrapper[4813]: I1206 16:14:56.952000 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nbw6k" event={"ID":"c8332abd-e13b-4587-831b-1330e18de573","Type":"ContainerStarted","Data":"a0993edeab00640ace43a44b340585f46bcf65baae780a0610d2974361739215"} Dec 06 16:14:56 crc kubenswrapper[4813]: I1206 16:14:56.978302 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nbw6k" podStartSLOduration=2.339995941 podStartE2EDuration="2.978256827s" podCreationTimestamp="2025-12-06 16:14:54 +0000 UTC" firstStartedPulling="2025-12-06 16:14:55.342639165 +0000 UTC m=+1735.233518751" lastFinishedPulling="2025-12-06 16:14:55.980900021 +0000 UTC m=+1735.871779637" observedRunningTime="2025-12-06 16:14:56.975510829 +0000 UTC m=+1736.866390435" watchObservedRunningTime="2025-12-06 16:14:56.978256827 +0000 UTC m=+1736.869136433" Dec 06 16:15:00 crc kubenswrapper[4813]: I1206 16:15:00.142753 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417295-wldtx"] Dec 06 16:15:00 crc kubenswrapper[4813]: I1206 16:15:00.145496 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417295-wldtx" Dec 06 16:15:00 crc kubenswrapper[4813]: I1206 16:15:00.147697 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 16:15:00 crc kubenswrapper[4813]: I1206 16:15:00.147739 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 16:15:00 crc kubenswrapper[4813]: I1206 16:15:00.165912 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417295-wldtx"] Dec 06 16:15:00 crc kubenswrapper[4813]: I1206 16:15:00.304189 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5d9f094-18ad-43ec-9f34-fa12baab1366-secret-volume\") pod \"collect-profiles-29417295-wldtx\" (UID: \"a5d9f094-18ad-43ec-9f34-fa12baab1366\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417295-wldtx" Dec 06 16:15:00 crc kubenswrapper[4813]: I1206 16:15:00.304234 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5d9f094-18ad-43ec-9f34-fa12baab1366-config-volume\") pod \"collect-profiles-29417295-wldtx\" (UID: \"a5d9f094-18ad-43ec-9f34-fa12baab1366\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417295-wldtx" Dec 06 16:15:00 crc kubenswrapper[4813]: I1206 16:15:00.304355 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzhwf\" (UniqueName: \"kubernetes.io/projected/a5d9f094-18ad-43ec-9f34-fa12baab1366-kube-api-access-jzhwf\") pod \"collect-profiles-29417295-wldtx\" (UID: \"a5d9f094-18ad-43ec-9f34-fa12baab1366\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417295-wldtx" Dec 06 16:15:00 crc kubenswrapper[4813]: I1206 16:15:00.406481 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5d9f094-18ad-43ec-9f34-fa12baab1366-secret-volume\") pod \"collect-profiles-29417295-wldtx\" (UID: \"a5d9f094-18ad-43ec-9f34-fa12baab1366\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417295-wldtx" Dec 06 16:15:00 crc kubenswrapper[4813]: I1206 16:15:00.406541 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5d9f094-18ad-43ec-9f34-fa12baab1366-config-volume\") pod \"collect-profiles-29417295-wldtx\" (UID: \"a5d9f094-18ad-43ec-9f34-fa12baab1366\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417295-wldtx" Dec 06 16:15:00 crc kubenswrapper[4813]: I1206 16:15:00.406610 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzhwf\" (UniqueName: \"kubernetes.io/projected/a5d9f094-18ad-43ec-9f34-fa12baab1366-kube-api-access-jzhwf\") pod \"collect-profiles-29417295-wldtx\" (UID: \"a5d9f094-18ad-43ec-9f34-fa12baab1366\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417295-wldtx" Dec 06 16:15:00 crc kubenswrapper[4813]: I1206 16:15:00.407757 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5d9f094-18ad-43ec-9f34-fa12baab1366-config-volume\") pod \"collect-profiles-29417295-wldtx\" (UID: \"a5d9f094-18ad-43ec-9f34-fa12baab1366\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417295-wldtx" Dec 06 16:15:00 crc kubenswrapper[4813]: I1206 16:15:00.421727 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5d9f094-18ad-43ec-9f34-fa12baab1366-secret-volume\") pod \"collect-profiles-29417295-wldtx\" (UID: \"a5d9f094-18ad-43ec-9f34-fa12baab1366\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417295-wldtx" Dec 06 16:15:00 crc kubenswrapper[4813]: I1206 16:15:00.426315 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzhwf\" (UniqueName: \"kubernetes.io/projected/a5d9f094-18ad-43ec-9f34-fa12baab1366-kube-api-access-jzhwf\") pod \"collect-profiles-29417295-wldtx\" (UID: \"a5d9f094-18ad-43ec-9f34-fa12baab1366\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417295-wldtx" Dec 06 16:15:00 crc kubenswrapper[4813]: I1206 16:15:00.473867 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 16:15:00 crc kubenswrapper[4813]: I1206 16:15:00.482542 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417295-wldtx" Dec 06 16:15:00 crc kubenswrapper[4813]: I1206 16:15:00.736228 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417295-wldtx"] Dec 06 16:15:00 crc kubenswrapper[4813]: I1206 16:15:00.993640 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417295-wldtx" event={"ID":"a5d9f094-18ad-43ec-9f34-fa12baab1366","Type":"ContainerStarted","Data":"b73b4ec3f29b81ef63119db5f8ffabb5299042a2858d2f571ca9ecd7e2d8425f"} Dec 06 16:15:00 crc kubenswrapper[4813]: I1206 16:15:00.993721 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417295-wldtx" event={"ID":"a5d9f094-18ad-43ec-9f34-fa12baab1366","Type":"ContainerStarted","Data":"7d537e481aecf5a101cee5b30a571243bfbecfcf801b0b3cbe7f7b0e8e09aa93"} Dec 06 16:15:01 crc kubenswrapper[4813]: I1206 16:15:01.014735 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29417295-wldtx" podStartSLOduration=1.014719318 podStartE2EDuration="1.014719318s" podCreationTimestamp="2025-12-06 16:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:15:01.006637169 +0000 UTC m=+1740.897516745" watchObservedRunningTime="2025-12-06 16:15:01.014719318 +0000 UTC m=+1740.905598894" Dec 06 16:15:01 crc kubenswrapper[4813]: I1206 16:15:01.047090 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-pn5s8"] Dec 06 16:15:01 crc kubenswrapper[4813]: I1206 16:15:01.055901 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-pn5s8"] Dec 06 16:15:02 crc kubenswrapper[4813]: I1206 16:15:02.005157 4813 generic.go:334] "Generic (PLEG): container finished" podID="a5d9f094-18ad-43ec-9f34-fa12baab1366" containerID="b73b4ec3f29b81ef63119db5f8ffabb5299042a2858d2f571ca9ecd7e2d8425f" exitCode=0 Dec 06 16:15:02 crc kubenswrapper[4813]: I1206 16:15:02.005368 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417295-wldtx" event={"ID":"a5d9f094-18ad-43ec-9f34-fa12baab1366","Type":"ContainerDied","Data":"b73b4ec3f29b81ef63119db5f8ffabb5299042a2858d2f571ca9ecd7e2d8425f"} Dec 06 16:15:02 crc kubenswrapper[4813]: I1206 16:15:02.490986 4813 scope.go:117] "RemoveContainer" containerID="dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412" Dec 06 16:15:02 crc kubenswrapper[4813]: E1206 16:15:02.491894 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:15:02 crc kubenswrapper[4813]: I1206 16:15:02.500876 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ff658cc-71bb-4df5-8ba8-1c0979e03d10" path="/var/lib/kubelet/pods/2ff658cc-71bb-4df5-8ba8-1c0979e03d10/volumes" Dec 06 16:15:03 crc kubenswrapper[4813]: I1206 16:15:03.395398 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417295-wldtx" Dec 06 16:15:03 crc kubenswrapper[4813]: I1206 16:15:03.468658 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5d9f094-18ad-43ec-9f34-fa12baab1366-config-volume\") pod \"a5d9f094-18ad-43ec-9f34-fa12baab1366\" (UID: \"a5d9f094-18ad-43ec-9f34-fa12baab1366\") " Dec 06 16:15:03 crc kubenswrapper[4813]: I1206 16:15:03.468853 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzhwf\" (UniqueName: \"kubernetes.io/projected/a5d9f094-18ad-43ec-9f34-fa12baab1366-kube-api-access-jzhwf\") pod \"a5d9f094-18ad-43ec-9f34-fa12baab1366\" (UID: \"a5d9f094-18ad-43ec-9f34-fa12baab1366\") " Dec 06 16:15:03 crc kubenswrapper[4813]: I1206 16:15:03.468971 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5d9f094-18ad-43ec-9f34-fa12baab1366-secret-volume\") pod \"a5d9f094-18ad-43ec-9f34-fa12baab1366\" (UID: \"a5d9f094-18ad-43ec-9f34-fa12baab1366\") " Dec 06 16:15:03 crc kubenswrapper[4813]: I1206 16:15:03.469242 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5d9f094-18ad-43ec-9f34-fa12baab1366-config-volume" (OuterVolumeSpecName: "config-volume") pod "a5d9f094-18ad-43ec-9f34-fa12baab1366" (UID: "a5d9f094-18ad-43ec-9f34-fa12baab1366"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:15:03 crc kubenswrapper[4813]: I1206 16:15:03.469539 4813 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5d9f094-18ad-43ec-9f34-fa12baab1366-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 16:15:03 crc kubenswrapper[4813]: I1206 16:15:03.473951 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5d9f094-18ad-43ec-9f34-fa12baab1366-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a5d9f094-18ad-43ec-9f34-fa12baab1366" (UID: "a5d9f094-18ad-43ec-9f34-fa12baab1366"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:15:03 crc kubenswrapper[4813]: I1206 16:15:03.473956 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5d9f094-18ad-43ec-9f34-fa12baab1366-kube-api-access-jzhwf" (OuterVolumeSpecName: "kube-api-access-jzhwf") pod "a5d9f094-18ad-43ec-9f34-fa12baab1366" (UID: "a5d9f094-18ad-43ec-9f34-fa12baab1366"). InnerVolumeSpecName "kube-api-access-jzhwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:15:03 crc kubenswrapper[4813]: I1206 16:15:03.570956 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzhwf\" (UniqueName: \"kubernetes.io/projected/a5d9f094-18ad-43ec-9f34-fa12baab1366-kube-api-access-jzhwf\") on node \"crc\" DevicePath \"\"" Dec 06 16:15:03 crc kubenswrapper[4813]: I1206 16:15:03.570981 4813 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5d9f094-18ad-43ec-9f34-fa12baab1366-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 16:15:04 crc kubenswrapper[4813]: I1206 16:15:04.029934 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417295-wldtx" event={"ID":"a5d9f094-18ad-43ec-9f34-fa12baab1366","Type":"ContainerDied","Data":"7d537e481aecf5a101cee5b30a571243bfbecfcf801b0b3cbe7f7b0e8e09aa93"} Dec 06 16:15:04 crc kubenswrapper[4813]: I1206 16:15:04.029978 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d537e481aecf5a101cee5b30a571243bfbecfcf801b0b3cbe7f7b0e8e09aa93" Dec 06 16:15:04 crc kubenswrapper[4813]: I1206 16:15:04.030033 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417295-wldtx" Dec 06 16:15:17 crc kubenswrapper[4813]: I1206 16:15:17.487129 4813 scope.go:117] "RemoveContainer" containerID="dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412" Dec 06 16:15:17 crc kubenswrapper[4813]: E1206 16:15:17.488362 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:15:24 crc kubenswrapper[4813]: I1206 16:15:24.012633 4813 scope.go:117] "RemoveContainer" containerID="2b018776e90533759f52261401f35f1e5b333dd629be589509b2457f0a31dd8e" Dec 06 16:15:24 crc kubenswrapper[4813]: I1206 16:15:24.037775 4813 scope.go:117] "RemoveContainer" containerID="9eab219b857b50d7f456aa7914cb1df32c1e15bec7ae7688ec8af024f141900f" Dec 06 16:15:24 crc kubenswrapper[4813]: I1206 16:15:24.099841 4813 scope.go:117] "RemoveContainer" containerID="4cdc7c62d685ecd333aa00daac5609ba7f02b43330e726059f3482413810fd1e" Dec 06 16:15:24 crc kubenswrapper[4813]: I1206 16:15:24.130395 4813 scope.go:117] "RemoveContainer" containerID="adda738648174e8dcbda04a4c13cbb2f263e4a82465ea6f49b4c3082ed252d21" Dec 06 16:15:24 crc kubenswrapper[4813]: I1206 16:15:24.178533 4813 scope.go:117] "RemoveContainer" containerID="c8f9099fcb752af009bd425b0fbf2b15bdbfb56e518110eec0f2d4b689301885" Dec 06 16:15:24 crc kubenswrapper[4813]: I1206 16:15:24.222688 4813 scope.go:117] "RemoveContainer" containerID="75bb59db87ccb27e3651904a042c755b9be081706b6842c2bd2ad04120ed6287" Dec 06 16:15:24 crc kubenswrapper[4813]: I1206 16:15:24.245727 4813 scope.go:117] "RemoveContainer" containerID="74d957793a0594fe543c66f2360a457565a7f5aef669690ca0c52d34fe724784" Dec 06 16:15:26 crc kubenswrapper[4813]: I1206 16:15:26.041382 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-xwdwl"] Dec 06 16:15:26 crc kubenswrapper[4813]: I1206 16:15:26.059545 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-xwdwl"] Dec 06 16:15:26 crc kubenswrapper[4813]: I1206 16:15:26.503205 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fc8a5f4-00b7-434b-b375-74d5a95e3fd1" path="/var/lib/kubelet/pods/8fc8a5f4-00b7-434b-b375-74d5a95e3fd1/volumes" Dec 06 16:15:31 crc kubenswrapper[4813]: I1206 16:15:31.486538 4813 scope.go:117] "RemoveContainer" containerID="dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412" Dec 06 16:15:31 crc kubenswrapper[4813]: E1206 16:15:31.487295 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:15:32 crc kubenswrapper[4813]: I1206 16:15:32.053203 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gv4hp"] Dec 06 16:15:32 crc kubenswrapper[4813]: I1206 16:15:32.070479 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gv4hp"] Dec 06 16:15:32 crc kubenswrapper[4813]: I1206 16:15:32.496877 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d9f1a15-0cf8-45e2-b4f4-d15d257d7009" path="/var/lib/kubelet/pods/4d9f1a15-0cf8-45e2-b4f4-d15d257d7009/volumes" Dec 06 16:15:38 crc kubenswrapper[4813]: I1206 16:15:38.438238 4813 generic.go:334] "Generic (PLEG): container finished" podID="c8332abd-e13b-4587-831b-1330e18de573" containerID="a0993edeab00640ace43a44b340585f46bcf65baae780a0610d2974361739215" exitCode=0 Dec 06 16:15:38 crc kubenswrapper[4813]: I1206 16:15:38.438461 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nbw6k" event={"ID":"c8332abd-e13b-4587-831b-1330e18de573","Type":"ContainerDied","Data":"a0993edeab00640ace43a44b340585f46bcf65baae780a0610d2974361739215"} Dec 06 16:15:39 crc kubenswrapper[4813]: I1206 16:15:39.868954 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nbw6k" Dec 06 16:15:39 crc kubenswrapper[4813]: I1206 16:15:39.938073 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wttqb\" (UniqueName: \"kubernetes.io/projected/c8332abd-e13b-4587-831b-1330e18de573-kube-api-access-wttqb\") pod \"c8332abd-e13b-4587-831b-1330e18de573\" (UID: \"c8332abd-e13b-4587-831b-1330e18de573\") " Dec 06 16:15:39 crc kubenswrapper[4813]: I1206 16:15:39.938207 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c8332abd-e13b-4587-831b-1330e18de573-inventory\") pod \"c8332abd-e13b-4587-831b-1330e18de573\" (UID: \"c8332abd-e13b-4587-831b-1330e18de573\") " Dec 06 16:15:39 crc kubenswrapper[4813]: I1206 16:15:39.938242 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c8332abd-e13b-4587-831b-1330e18de573-ssh-key\") pod \"c8332abd-e13b-4587-831b-1330e18de573\" (UID: \"c8332abd-e13b-4587-831b-1330e18de573\") " Dec 06 16:15:39 crc kubenswrapper[4813]: I1206 16:15:39.948337 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8332abd-e13b-4587-831b-1330e18de573-kube-api-access-wttqb" (OuterVolumeSpecName: "kube-api-access-wttqb") pod "c8332abd-e13b-4587-831b-1330e18de573" (UID: "c8332abd-e13b-4587-831b-1330e18de573"). InnerVolumeSpecName "kube-api-access-wttqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:15:39 crc kubenswrapper[4813]: I1206 16:15:39.966771 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8332abd-e13b-4587-831b-1330e18de573-inventory" (OuterVolumeSpecName: "inventory") pod "c8332abd-e13b-4587-831b-1330e18de573" (UID: "c8332abd-e13b-4587-831b-1330e18de573"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:15:39 crc kubenswrapper[4813]: I1206 16:15:39.968605 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8332abd-e13b-4587-831b-1330e18de573-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c8332abd-e13b-4587-831b-1330e18de573" (UID: "c8332abd-e13b-4587-831b-1330e18de573"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:15:40 crc kubenswrapper[4813]: I1206 16:15:40.041211 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wttqb\" (UniqueName: \"kubernetes.io/projected/c8332abd-e13b-4587-831b-1330e18de573-kube-api-access-wttqb\") on node \"crc\" DevicePath \"\"" Dec 06 16:15:40 crc kubenswrapper[4813]: I1206 16:15:40.041252 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c8332abd-e13b-4587-831b-1330e18de573-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 16:15:40 crc kubenswrapper[4813]: I1206 16:15:40.041281 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c8332abd-e13b-4587-831b-1330e18de573-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 16:15:40 crc kubenswrapper[4813]: I1206 16:15:40.460126 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nbw6k" event={"ID":"c8332abd-e13b-4587-831b-1330e18de573","Type":"ContainerDied","Data":"39e00e030492b64bb72bac3a48f46d57f748455521b327a65577cbfbe0e26346"} Dec 06 16:15:40 crc kubenswrapper[4813]: I1206 16:15:40.460185 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nbw6k" Dec 06 16:15:40 crc kubenswrapper[4813]: I1206 16:15:40.460194 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39e00e030492b64bb72bac3a48f46d57f748455521b327a65577cbfbe0e26346" Dec 06 16:15:40 crc kubenswrapper[4813]: I1206 16:15:40.584768 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl"] Dec 06 16:15:40 crc kubenswrapper[4813]: E1206 16:15:40.585155 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5d9f094-18ad-43ec-9f34-fa12baab1366" containerName="collect-profiles" Dec 06 16:15:40 crc kubenswrapper[4813]: I1206 16:15:40.585170 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5d9f094-18ad-43ec-9f34-fa12baab1366" containerName="collect-profiles" Dec 06 16:15:40 crc kubenswrapper[4813]: E1206 16:15:40.585191 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8332abd-e13b-4587-831b-1330e18de573" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 06 16:15:40 crc kubenswrapper[4813]: I1206 16:15:40.585200 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8332abd-e13b-4587-831b-1330e18de573" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 06 16:15:40 crc kubenswrapper[4813]: I1206 16:15:40.585435 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8332abd-e13b-4587-831b-1330e18de573" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 06 16:15:40 crc kubenswrapper[4813]: I1206 16:15:40.585455 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5d9f094-18ad-43ec-9f34-fa12baab1366" containerName="collect-profiles" Dec 06 16:15:40 crc kubenswrapper[4813]: I1206 16:15:40.586138 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl" Dec 06 16:15:40 crc kubenswrapper[4813]: I1206 16:15:40.589674 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 16:15:40 crc kubenswrapper[4813]: I1206 16:15:40.591409 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jvkhf" Dec 06 16:15:40 crc kubenswrapper[4813]: I1206 16:15:40.591532 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 16:15:40 crc kubenswrapper[4813]: I1206 16:15:40.591662 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 16:15:40 crc kubenswrapper[4813]: I1206 16:15:40.599424 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl"] Dec 06 16:15:40 crc kubenswrapper[4813]: I1206 16:15:40.654795 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6f525ab6-f23e-4cc9-a843-7a601896c864-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl\" (UID: \"6f525ab6-f23e-4cc9-a843-7a601896c864\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl" Dec 06 16:15:40 crc kubenswrapper[4813]: I1206 16:15:40.654863 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfxqz\" (UniqueName: \"kubernetes.io/projected/6f525ab6-f23e-4cc9-a843-7a601896c864-kube-api-access-dfxqz\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl\" (UID: \"6f525ab6-f23e-4cc9-a843-7a601896c864\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl" Dec 06 16:15:40 crc kubenswrapper[4813]: I1206 16:15:40.655103 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f525ab6-f23e-4cc9-a843-7a601896c864-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl\" (UID: \"6f525ab6-f23e-4cc9-a843-7a601896c864\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl" Dec 06 16:15:40 crc kubenswrapper[4813]: I1206 16:15:40.757502 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f525ab6-f23e-4cc9-a843-7a601896c864-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl\" (UID: \"6f525ab6-f23e-4cc9-a843-7a601896c864\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl" Dec 06 16:15:40 crc kubenswrapper[4813]: I1206 16:15:40.757647 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6f525ab6-f23e-4cc9-a843-7a601896c864-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl\" (UID: \"6f525ab6-f23e-4cc9-a843-7a601896c864\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl" Dec 06 16:15:40 crc kubenswrapper[4813]: I1206 16:15:40.757707 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfxqz\" (UniqueName: \"kubernetes.io/projected/6f525ab6-f23e-4cc9-a843-7a601896c864-kube-api-access-dfxqz\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl\" (UID: \"6f525ab6-f23e-4cc9-a843-7a601896c864\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl" Dec 06 16:15:40 crc kubenswrapper[4813]: I1206 16:15:40.764715 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f525ab6-f23e-4cc9-a843-7a601896c864-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl\" (UID: \"6f525ab6-f23e-4cc9-a843-7a601896c864\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl" Dec 06 16:15:40 crc kubenswrapper[4813]: I1206 16:15:40.768003 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6f525ab6-f23e-4cc9-a843-7a601896c864-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl\" (UID: \"6f525ab6-f23e-4cc9-a843-7a601896c864\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl" Dec 06 16:15:40 crc kubenswrapper[4813]: I1206 16:15:40.783765 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfxqz\" (UniqueName: \"kubernetes.io/projected/6f525ab6-f23e-4cc9-a843-7a601896c864-kube-api-access-dfxqz\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl\" (UID: \"6f525ab6-f23e-4cc9-a843-7a601896c864\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl" Dec 06 16:15:40 crc kubenswrapper[4813]: I1206 16:15:40.914604 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl" Dec 06 16:15:41 crc kubenswrapper[4813]: I1206 16:15:41.433631 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl"] Dec 06 16:15:41 crc kubenswrapper[4813]: I1206 16:15:41.470041 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl" event={"ID":"6f525ab6-f23e-4cc9-a843-7a601896c864","Type":"ContainerStarted","Data":"e5daa86a51af30e8f6f6dddcaeaaf42ec700e257cb30c958c46a93bc1df236c0"} Dec 06 16:15:42 crc kubenswrapper[4813]: I1206 16:15:42.485786 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl" event={"ID":"6f525ab6-f23e-4cc9-a843-7a601896c864","Type":"ContainerStarted","Data":"e2e3f352464bc6e4ccf4aae6a9825b6be007bf7d20e9d540f92e5802034b6257"} Dec 06 16:15:42 crc kubenswrapper[4813]: I1206 16:15:42.507571 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl" podStartSLOduration=2.070663886 podStartE2EDuration="2.507553079s" podCreationTimestamp="2025-12-06 16:15:40 +0000 UTC" firstStartedPulling="2025-12-06 16:15:41.447427969 +0000 UTC m=+1781.338307545" lastFinishedPulling="2025-12-06 16:15:41.884317132 +0000 UTC m=+1781.775196738" observedRunningTime="2025-12-06 16:15:42.500751131 +0000 UTC m=+1782.391630707" watchObservedRunningTime="2025-12-06 16:15:42.507553079 +0000 UTC m=+1782.398432655" Dec 06 16:15:45 crc kubenswrapper[4813]: I1206 16:15:45.487315 4813 scope.go:117] "RemoveContainer" containerID="dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412" Dec 06 16:15:45 crc kubenswrapper[4813]: E1206 16:15:45.487828 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:16:00 crc kubenswrapper[4813]: I1206 16:16:00.492187 4813 scope.go:117] "RemoveContainer" containerID="dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412" Dec 06 16:16:00 crc kubenswrapper[4813]: E1206 16:16:00.495669 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:16:10 crc kubenswrapper[4813]: I1206 16:16:10.056305 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-nq29n"] Dec 06 16:16:10 crc kubenswrapper[4813]: I1206 16:16:10.070442 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-nq29n"] Dec 06 16:16:10 crc kubenswrapper[4813]: I1206 16:16:10.501021 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b77b30e-c026-408e-a869-0bdc2c1d650a" path="/var/lib/kubelet/pods/1b77b30e-c026-408e-a869-0bdc2c1d650a/volumes" Dec 06 16:16:11 crc kubenswrapper[4813]: I1206 16:16:11.486670 4813 scope.go:117] "RemoveContainer" containerID="dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412" Dec 06 16:16:11 crc kubenswrapper[4813]: E1206 16:16:11.487224 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:16:24 crc kubenswrapper[4813]: I1206 16:16:24.399109 4813 scope.go:117] "RemoveContainer" containerID="987af28cdd40354bf772b307c353116f8cff0959e35004737b938db61dea3de7" Dec 06 16:16:24 crc kubenswrapper[4813]: I1206 16:16:24.471635 4813 scope.go:117] "RemoveContainer" containerID="bd8568faea71fd736f6a65c700cae646473782de03144171c2d8ade2038b5179" Dec 06 16:16:24 crc kubenswrapper[4813]: I1206 16:16:24.516101 4813 scope.go:117] "RemoveContainer" containerID="a4537509eb048af72dd28535e54e3ecd586c8ede261a2a308f156c700921caa1" Dec 06 16:16:26 crc kubenswrapper[4813]: I1206 16:16:26.487403 4813 scope.go:117] "RemoveContainer" containerID="dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412" Dec 06 16:16:26 crc kubenswrapper[4813]: E1206 16:16:26.488032 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:16:40 crc kubenswrapper[4813]: I1206 16:16:40.494728 4813 scope.go:117] "RemoveContainer" containerID="dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412" Dec 06 16:16:40 crc kubenswrapper[4813]: E1206 16:16:40.495771 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:16:42 crc kubenswrapper[4813]: I1206 16:16:42.070496 4813 generic.go:334] "Generic (PLEG): container finished" podID="6f525ab6-f23e-4cc9-a843-7a601896c864" containerID="e2e3f352464bc6e4ccf4aae6a9825b6be007bf7d20e9d540f92e5802034b6257" exitCode=0 Dec 06 16:16:42 crc kubenswrapper[4813]: I1206 16:16:42.070723 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl" event={"ID":"6f525ab6-f23e-4cc9-a843-7a601896c864","Type":"ContainerDied","Data":"e2e3f352464bc6e4ccf4aae6a9825b6be007bf7d20e9d540f92e5802034b6257"} Dec 06 16:16:43 crc kubenswrapper[4813]: I1206 16:16:43.537927 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl" Dec 06 16:16:43 crc kubenswrapper[4813]: I1206 16:16:43.680754 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfxqz\" (UniqueName: \"kubernetes.io/projected/6f525ab6-f23e-4cc9-a843-7a601896c864-kube-api-access-dfxqz\") pod \"6f525ab6-f23e-4cc9-a843-7a601896c864\" (UID: \"6f525ab6-f23e-4cc9-a843-7a601896c864\") " Dec 06 16:16:43 crc kubenswrapper[4813]: I1206 16:16:43.680809 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6f525ab6-f23e-4cc9-a843-7a601896c864-inventory\") pod \"6f525ab6-f23e-4cc9-a843-7a601896c864\" (UID: \"6f525ab6-f23e-4cc9-a843-7a601896c864\") " Dec 06 16:16:43 crc kubenswrapper[4813]: I1206 16:16:43.680837 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f525ab6-f23e-4cc9-a843-7a601896c864-ssh-key\") pod \"6f525ab6-f23e-4cc9-a843-7a601896c864\" (UID: \"6f525ab6-f23e-4cc9-a843-7a601896c864\") " Dec 06 16:16:43 crc kubenswrapper[4813]: I1206 16:16:43.691509 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f525ab6-f23e-4cc9-a843-7a601896c864-kube-api-access-dfxqz" (OuterVolumeSpecName: "kube-api-access-dfxqz") pod "6f525ab6-f23e-4cc9-a843-7a601896c864" (UID: "6f525ab6-f23e-4cc9-a843-7a601896c864"). InnerVolumeSpecName "kube-api-access-dfxqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:16:43 crc kubenswrapper[4813]: I1206 16:16:43.715342 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f525ab6-f23e-4cc9-a843-7a601896c864-inventory" (OuterVolumeSpecName: "inventory") pod "6f525ab6-f23e-4cc9-a843-7a601896c864" (UID: "6f525ab6-f23e-4cc9-a843-7a601896c864"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:16:43 crc kubenswrapper[4813]: I1206 16:16:43.727308 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f525ab6-f23e-4cc9-a843-7a601896c864-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6f525ab6-f23e-4cc9-a843-7a601896c864" (UID: "6f525ab6-f23e-4cc9-a843-7a601896c864"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:16:43 crc kubenswrapper[4813]: I1206 16:16:43.783581 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f525ab6-f23e-4cc9-a843-7a601896c864-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 16:16:43 crc kubenswrapper[4813]: I1206 16:16:43.783713 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfxqz\" (UniqueName: \"kubernetes.io/projected/6f525ab6-f23e-4cc9-a843-7a601896c864-kube-api-access-dfxqz\") on node \"crc\" DevicePath \"\"" Dec 06 16:16:43 crc kubenswrapper[4813]: I1206 16:16:43.783772 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6f525ab6-f23e-4cc9-a843-7a601896c864-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 16:16:44 crc kubenswrapper[4813]: I1206 16:16:44.091085 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl" event={"ID":"6f525ab6-f23e-4cc9-a843-7a601896c864","Type":"ContainerDied","Data":"e5daa86a51af30e8f6f6dddcaeaaf42ec700e257cb30c958c46a93bc1df236c0"} Dec 06 16:16:44 crc kubenswrapper[4813]: I1206 16:16:44.091119 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5daa86a51af30e8f6f6dddcaeaaf42ec700e257cb30c958c46a93bc1df236c0" Dec 06 16:16:44 crc kubenswrapper[4813]: I1206 16:16:44.091183 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl" Dec 06 16:16:44 crc kubenswrapper[4813]: I1206 16:16:44.179791 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-nj5n7"] Dec 06 16:16:44 crc kubenswrapper[4813]: E1206 16:16:44.180220 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f525ab6-f23e-4cc9-a843-7a601896c864" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 06 16:16:44 crc kubenswrapper[4813]: I1206 16:16:44.180241 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f525ab6-f23e-4cc9-a843-7a601896c864" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 06 16:16:44 crc kubenswrapper[4813]: I1206 16:16:44.180525 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f525ab6-f23e-4cc9-a843-7a601896c864" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 06 16:16:44 crc kubenswrapper[4813]: I1206 16:16:44.181251 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-nj5n7" Dec 06 16:16:44 crc kubenswrapper[4813]: I1206 16:16:44.185002 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 16:16:44 crc kubenswrapper[4813]: I1206 16:16:44.185337 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 16:16:44 crc kubenswrapper[4813]: I1206 16:16:44.185465 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jvkhf" Dec 06 16:16:44 crc kubenswrapper[4813]: I1206 16:16:44.185593 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 16:16:44 crc kubenswrapper[4813]: I1206 16:16:44.198199 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-nj5n7"] Dec 06 16:16:44 crc kubenswrapper[4813]: I1206 16:16:44.292562 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/41e561aa-3484-4946-a8d9-174ea9a22a45-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-nj5n7\" (UID: \"41e561aa-3484-4946-a8d9-174ea9a22a45\") " pod="openstack/ssh-known-hosts-edpm-deployment-nj5n7" Dec 06 16:16:44 crc kubenswrapper[4813]: I1206 16:16:44.292641 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/41e561aa-3484-4946-a8d9-174ea9a22a45-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-nj5n7\" (UID: \"41e561aa-3484-4946-a8d9-174ea9a22a45\") " pod="openstack/ssh-known-hosts-edpm-deployment-nj5n7" Dec 06 16:16:44 crc kubenswrapper[4813]: I1206 16:16:44.292700 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm7vk\" (UniqueName: \"kubernetes.io/projected/41e561aa-3484-4946-a8d9-174ea9a22a45-kube-api-access-lm7vk\") pod \"ssh-known-hosts-edpm-deployment-nj5n7\" (UID: \"41e561aa-3484-4946-a8d9-174ea9a22a45\") " pod="openstack/ssh-known-hosts-edpm-deployment-nj5n7" Dec 06 16:16:44 crc kubenswrapper[4813]: I1206 16:16:44.394050 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/41e561aa-3484-4946-a8d9-174ea9a22a45-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-nj5n7\" (UID: \"41e561aa-3484-4946-a8d9-174ea9a22a45\") " pod="openstack/ssh-known-hosts-edpm-deployment-nj5n7" Dec 06 16:16:44 crc kubenswrapper[4813]: I1206 16:16:44.394112 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/41e561aa-3484-4946-a8d9-174ea9a22a45-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-nj5n7\" (UID: \"41e561aa-3484-4946-a8d9-174ea9a22a45\") " pod="openstack/ssh-known-hosts-edpm-deployment-nj5n7" Dec 06 16:16:44 crc kubenswrapper[4813]: I1206 16:16:44.394142 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm7vk\" (UniqueName: \"kubernetes.io/projected/41e561aa-3484-4946-a8d9-174ea9a22a45-kube-api-access-lm7vk\") pod \"ssh-known-hosts-edpm-deployment-nj5n7\" (UID: \"41e561aa-3484-4946-a8d9-174ea9a22a45\") " pod="openstack/ssh-known-hosts-edpm-deployment-nj5n7" Dec 06 16:16:44 crc kubenswrapper[4813]: I1206 16:16:44.399746 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/41e561aa-3484-4946-a8d9-174ea9a22a45-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-nj5n7\" (UID: \"41e561aa-3484-4946-a8d9-174ea9a22a45\") " pod="openstack/ssh-known-hosts-edpm-deployment-nj5n7" Dec 06 16:16:44 crc kubenswrapper[4813]: I1206 16:16:44.402866 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/41e561aa-3484-4946-a8d9-174ea9a22a45-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-nj5n7\" (UID: \"41e561aa-3484-4946-a8d9-174ea9a22a45\") " pod="openstack/ssh-known-hosts-edpm-deployment-nj5n7" Dec 06 16:16:44 crc kubenswrapper[4813]: I1206 16:16:44.414590 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm7vk\" (UniqueName: \"kubernetes.io/projected/41e561aa-3484-4946-a8d9-174ea9a22a45-kube-api-access-lm7vk\") pod \"ssh-known-hosts-edpm-deployment-nj5n7\" (UID: \"41e561aa-3484-4946-a8d9-174ea9a22a45\") " pod="openstack/ssh-known-hosts-edpm-deployment-nj5n7" Dec 06 16:16:44 crc kubenswrapper[4813]: I1206 16:16:44.503198 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-nj5n7" Dec 06 16:16:45 crc kubenswrapper[4813]: I1206 16:16:45.044868 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-nj5n7"] Dec 06 16:16:45 crc kubenswrapper[4813]: I1206 16:16:45.063898 4813 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 16:16:45 crc kubenswrapper[4813]: I1206 16:16:45.109186 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-nj5n7" event={"ID":"41e561aa-3484-4946-a8d9-174ea9a22a45","Type":"ContainerStarted","Data":"e81cbd32ad269641c7d7cf294793dc4f46af7038a03e1146fb358b57794cde1f"} Dec 06 16:16:46 crc kubenswrapper[4813]: I1206 16:16:46.117314 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-nj5n7" event={"ID":"41e561aa-3484-4946-a8d9-174ea9a22a45","Type":"ContainerStarted","Data":"80f26ccd65959a2627592b6873d1329efa30110fdaacdbdd93bf2b83c4b8f3d5"} Dec 06 16:16:46 crc kubenswrapper[4813]: I1206 16:16:46.148698 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-nj5n7" podStartSLOduration=1.779188697 podStartE2EDuration="2.148677835s" podCreationTimestamp="2025-12-06 16:16:44 +0000 UTC" firstStartedPulling="2025-12-06 16:16:45.063405059 +0000 UTC m=+1844.954284635" lastFinishedPulling="2025-12-06 16:16:45.432894187 +0000 UTC m=+1845.323773773" observedRunningTime="2025-12-06 16:16:46.138027525 +0000 UTC m=+1846.028907101" watchObservedRunningTime="2025-12-06 16:16:46.148677835 +0000 UTC m=+1846.039557411" Dec 06 16:16:52 crc kubenswrapper[4813]: I1206 16:16:52.487248 4813 scope.go:117] "RemoveContainer" containerID="dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412" Dec 06 16:16:52 crc kubenswrapper[4813]: E1206 16:16:52.488422 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:16:54 crc kubenswrapper[4813]: I1206 16:16:54.329662 4813 generic.go:334] "Generic (PLEG): container finished" podID="41e561aa-3484-4946-a8d9-174ea9a22a45" containerID="80f26ccd65959a2627592b6873d1329efa30110fdaacdbdd93bf2b83c4b8f3d5" exitCode=0 Dec 06 16:16:54 crc kubenswrapper[4813]: I1206 16:16:54.329782 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-nj5n7" event={"ID":"41e561aa-3484-4946-a8d9-174ea9a22a45","Type":"ContainerDied","Data":"80f26ccd65959a2627592b6873d1329efa30110fdaacdbdd93bf2b83c4b8f3d5"} Dec 06 16:16:55 crc kubenswrapper[4813]: I1206 16:16:55.788784 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-nj5n7" Dec 06 16:16:55 crc kubenswrapper[4813]: I1206 16:16:55.902131 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/41e561aa-3484-4946-a8d9-174ea9a22a45-inventory-0\") pod \"41e561aa-3484-4946-a8d9-174ea9a22a45\" (UID: \"41e561aa-3484-4946-a8d9-174ea9a22a45\") " Dec 06 16:16:55 crc kubenswrapper[4813]: I1206 16:16:55.902522 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lm7vk\" (UniqueName: \"kubernetes.io/projected/41e561aa-3484-4946-a8d9-174ea9a22a45-kube-api-access-lm7vk\") pod \"41e561aa-3484-4946-a8d9-174ea9a22a45\" (UID: \"41e561aa-3484-4946-a8d9-174ea9a22a45\") " Dec 06 16:16:55 crc kubenswrapper[4813]: I1206 16:16:55.902599 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/41e561aa-3484-4946-a8d9-174ea9a22a45-ssh-key-openstack-edpm-ipam\") pod \"41e561aa-3484-4946-a8d9-174ea9a22a45\" (UID: \"41e561aa-3484-4946-a8d9-174ea9a22a45\") " Dec 06 16:16:55 crc kubenswrapper[4813]: I1206 16:16:55.907198 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41e561aa-3484-4946-a8d9-174ea9a22a45-kube-api-access-lm7vk" (OuterVolumeSpecName: "kube-api-access-lm7vk") pod "41e561aa-3484-4946-a8d9-174ea9a22a45" (UID: "41e561aa-3484-4946-a8d9-174ea9a22a45"). InnerVolumeSpecName "kube-api-access-lm7vk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:16:55 crc kubenswrapper[4813]: I1206 16:16:55.930856 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41e561aa-3484-4946-a8d9-174ea9a22a45-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "41e561aa-3484-4946-a8d9-174ea9a22a45" (UID: "41e561aa-3484-4946-a8d9-174ea9a22a45"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:16:55 crc kubenswrapper[4813]: I1206 16:16:55.947879 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41e561aa-3484-4946-a8d9-174ea9a22a45-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "41e561aa-3484-4946-a8d9-174ea9a22a45" (UID: "41e561aa-3484-4946-a8d9-174ea9a22a45"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.005560 4813 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/41e561aa-3484-4946-a8d9-174ea9a22a45-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.005611 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lm7vk\" (UniqueName: \"kubernetes.io/projected/41e561aa-3484-4946-a8d9-174ea9a22a45-kube-api-access-lm7vk\") on node \"crc\" DevicePath \"\"" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.005626 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/41e561aa-3484-4946-a8d9-174ea9a22a45-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.063502 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wc55n"] Dec 06 16:16:56 crc kubenswrapper[4813]: E1206 16:16:56.063880 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41e561aa-3484-4946-a8d9-174ea9a22a45" containerName="ssh-known-hosts-edpm-deployment" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.063891 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="41e561aa-3484-4946-a8d9-174ea9a22a45" containerName="ssh-known-hosts-edpm-deployment" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.064086 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="41e561aa-3484-4946-a8d9-174ea9a22a45" containerName="ssh-known-hosts-edpm-deployment" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.065817 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wc55n" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.087282 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wc55n"] Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.109952 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wjgn\" (UniqueName: \"kubernetes.io/projected/1afabcb9-e54f-4951-a71c-1dc4aff14b9b-kube-api-access-6wjgn\") pod \"certified-operators-wc55n\" (UID: \"1afabcb9-e54f-4951-a71c-1dc4aff14b9b\") " pod="openshift-marketplace/certified-operators-wc55n" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.110050 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1afabcb9-e54f-4951-a71c-1dc4aff14b9b-catalog-content\") pod \"certified-operators-wc55n\" (UID: \"1afabcb9-e54f-4951-a71c-1dc4aff14b9b\") " pod="openshift-marketplace/certified-operators-wc55n" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.110116 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1afabcb9-e54f-4951-a71c-1dc4aff14b9b-utilities\") pod \"certified-operators-wc55n\" (UID: \"1afabcb9-e54f-4951-a71c-1dc4aff14b9b\") " pod="openshift-marketplace/certified-operators-wc55n" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.211603 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wjgn\" (UniqueName: \"kubernetes.io/projected/1afabcb9-e54f-4951-a71c-1dc4aff14b9b-kube-api-access-6wjgn\") pod \"certified-operators-wc55n\" (UID: \"1afabcb9-e54f-4951-a71c-1dc4aff14b9b\") " pod="openshift-marketplace/certified-operators-wc55n" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.211675 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1afabcb9-e54f-4951-a71c-1dc4aff14b9b-catalog-content\") pod \"certified-operators-wc55n\" (UID: \"1afabcb9-e54f-4951-a71c-1dc4aff14b9b\") " pod="openshift-marketplace/certified-operators-wc55n" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.211709 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1afabcb9-e54f-4951-a71c-1dc4aff14b9b-utilities\") pod \"certified-operators-wc55n\" (UID: \"1afabcb9-e54f-4951-a71c-1dc4aff14b9b\") " pod="openshift-marketplace/certified-operators-wc55n" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.212243 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1afabcb9-e54f-4951-a71c-1dc4aff14b9b-utilities\") pod \"certified-operators-wc55n\" (UID: \"1afabcb9-e54f-4951-a71c-1dc4aff14b9b\") " pod="openshift-marketplace/certified-operators-wc55n" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.212349 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1afabcb9-e54f-4951-a71c-1dc4aff14b9b-catalog-content\") pod \"certified-operators-wc55n\" (UID: \"1afabcb9-e54f-4951-a71c-1dc4aff14b9b\") " pod="openshift-marketplace/certified-operators-wc55n" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.236240 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wjgn\" (UniqueName: \"kubernetes.io/projected/1afabcb9-e54f-4951-a71c-1dc4aff14b9b-kube-api-access-6wjgn\") pod \"certified-operators-wc55n\" (UID: \"1afabcb9-e54f-4951-a71c-1dc4aff14b9b\") " pod="openshift-marketplace/certified-operators-wc55n" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.353775 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-nj5n7" event={"ID":"41e561aa-3484-4946-a8d9-174ea9a22a45","Type":"ContainerDied","Data":"e81cbd32ad269641c7d7cf294793dc4f46af7038a03e1146fb358b57794cde1f"} Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.353820 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e81cbd32ad269641c7d7cf294793dc4f46af7038a03e1146fb358b57794cde1f" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.354331 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-nj5n7" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.385874 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wc55n" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.434003 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-h7rjq"] Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.435080 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h7rjq" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.443933 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.444106 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.444279 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jvkhf" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.444393 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.469254 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-h7rjq"] Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.516244 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/97035b28-6546-4c6b-b273-085f7ca5bb4f-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-h7rjq\" (UID: \"97035b28-6546-4c6b-b273-085f7ca5bb4f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h7rjq" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.516569 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/97035b28-6546-4c6b-b273-085f7ca5bb4f-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-h7rjq\" (UID: \"97035b28-6546-4c6b-b273-085f7ca5bb4f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h7rjq" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.516631 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5k4z4\" (UniqueName: \"kubernetes.io/projected/97035b28-6546-4c6b-b273-085f7ca5bb4f-kube-api-access-5k4z4\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-h7rjq\" (UID: \"97035b28-6546-4c6b-b273-085f7ca5bb4f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h7rjq" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.619743 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/97035b28-6546-4c6b-b273-085f7ca5bb4f-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-h7rjq\" (UID: \"97035b28-6546-4c6b-b273-085f7ca5bb4f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h7rjq" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.619874 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/97035b28-6546-4c6b-b273-085f7ca5bb4f-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-h7rjq\" (UID: \"97035b28-6546-4c6b-b273-085f7ca5bb4f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h7rjq" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.619933 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5k4z4\" (UniqueName: \"kubernetes.io/projected/97035b28-6546-4c6b-b273-085f7ca5bb4f-kube-api-access-5k4z4\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-h7rjq\" (UID: \"97035b28-6546-4c6b-b273-085f7ca5bb4f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h7rjq" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.633847 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/97035b28-6546-4c6b-b273-085f7ca5bb4f-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-h7rjq\" (UID: \"97035b28-6546-4c6b-b273-085f7ca5bb4f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h7rjq" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.644909 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5k4z4\" (UniqueName: \"kubernetes.io/projected/97035b28-6546-4c6b-b273-085f7ca5bb4f-kube-api-access-5k4z4\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-h7rjq\" (UID: \"97035b28-6546-4c6b-b273-085f7ca5bb4f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h7rjq" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.653315 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/97035b28-6546-4c6b-b273-085f7ca5bb4f-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-h7rjq\" (UID: \"97035b28-6546-4c6b-b273-085f7ca5bb4f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h7rjq" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.765924 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h7rjq" Dec 06 16:16:56 crc kubenswrapper[4813]: I1206 16:16:56.924045 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wc55n"] Dec 06 16:16:57 crc kubenswrapper[4813]: I1206 16:16:57.363967 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-h7rjq"] Dec 06 16:16:57 crc kubenswrapper[4813]: I1206 16:16:57.369935 4813 generic.go:334] "Generic (PLEG): container finished" podID="1afabcb9-e54f-4951-a71c-1dc4aff14b9b" containerID="5c69c86a391865121ffe076785973148cd92398374e59b2a0eeaeb12216499b6" exitCode=0 Dec 06 16:16:57 crc kubenswrapper[4813]: I1206 16:16:57.369973 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wc55n" event={"ID":"1afabcb9-e54f-4951-a71c-1dc4aff14b9b","Type":"ContainerDied","Data":"5c69c86a391865121ffe076785973148cd92398374e59b2a0eeaeb12216499b6"} Dec 06 16:16:57 crc kubenswrapper[4813]: I1206 16:16:57.369998 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wc55n" event={"ID":"1afabcb9-e54f-4951-a71c-1dc4aff14b9b","Type":"ContainerStarted","Data":"ce17840c0a664397163fc586c4c96760372e07b06cbeb9fee6555c86d40cccf0"} Dec 06 16:16:58 crc kubenswrapper[4813]: I1206 16:16:58.379212 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h7rjq" event={"ID":"97035b28-6546-4c6b-b273-085f7ca5bb4f","Type":"ContainerStarted","Data":"1d69d25070dc9c6f430d5b7ac80b7e7769f0024ef2d7b5c0c42c31965adc25d9"} Dec 06 16:16:58 crc kubenswrapper[4813]: I1206 16:16:58.379795 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h7rjq" event={"ID":"97035b28-6546-4c6b-b273-085f7ca5bb4f","Type":"ContainerStarted","Data":"4994fd7220ab0e21b7b620e7fd779fac80ab8da38b047704c077bbe0e97295d4"} Dec 06 16:16:58 crc kubenswrapper[4813]: I1206 16:16:58.383871 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wc55n" event={"ID":"1afabcb9-e54f-4951-a71c-1dc4aff14b9b","Type":"ContainerStarted","Data":"c922d922a9d7d740ce0dbd4c899b50d15093a211c05d31908e5a4f34ca7a8096"} Dec 06 16:16:58 crc kubenswrapper[4813]: I1206 16:16:58.402727 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h7rjq" podStartSLOduration=1.996262439 podStartE2EDuration="2.402710757s" podCreationTimestamp="2025-12-06 16:16:56 +0000 UTC" firstStartedPulling="2025-12-06 16:16:57.378660049 +0000 UTC m=+1857.269539635" lastFinishedPulling="2025-12-06 16:16:57.785108377 +0000 UTC m=+1857.675987953" observedRunningTime="2025-12-06 16:16:58.398724532 +0000 UTC m=+1858.289604118" watchObservedRunningTime="2025-12-06 16:16:58.402710757 +0000 UTC m=+1858.293590333" Dec 06 16:16:59 crc kubenswrapper[4813]: I1206 16:16:59.398096 4813 generic.go:334] "Generic (PLEG): container finished" podID="1afabcb9-e54f-4951-a71c-1dc4aff14b9b" containerID="c922d922a9d7d740ce0dbd4c899b50d15093a211c05d31908e5a4f34ca7a8096" exitCode=0 Dec 06 16:16:59 crc kubenswrapper[4813]: I1206 16:16:59.398170 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wc55n" event={"ID":"1afabcb9-e54f-4951-a71c-1dc4aff14b9b","Type":"ContainerDied","Data":"c922d922a9d7d740ce0dbd4c899b50d15093a211c05d31908e5a4f34ca7a8096"} Dec 06 16:17:00 crc kubenswrapper[4813]: I1206 16:17:00.412991 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wc55n" event={"ID":"1afabcb9-e54f-4951-a71c-1dc4aff14b9b","Type":"ContainerStarted","Data":"da84621a47fe6a1e220bf28a183562bc64d6584b538c1b7ac7f2c9bf385ad8c5"} Dec 06 16:17:00 crc kubenswrapper[4813]: I1206 16:17:00.441352 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wc55n" podStartSLOduration=1.683134811 podStartE2EDuration="4.441235572s" podCreationTimestamp="2025-12-06 16:16:56 +0000 UTC" firstStartedPulling="2025-12-06 16:16:57.373426102 +0000 UTC m=+1857.264305678" lastFinishedPulling="2025-12-06 16:17:00.131526853 +0000 UTC m=+1860.022406439" observedRunningTime="2025-12-06 16:17:00.431825785 +0000 UTC m=+1860.322705371" watchObservedRunningTime="2025-12-06 16:17:00.441235572 +0000 UTC m=+1860.332115158" Dec 06 16:17:04 crc kubenswrapper[4813]: I1206 16:17:04.487571 4813 scope.go:117] "RemoveContainer" containerID="dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412" Dec 06 16:17:04 crc kubenswrapper[4813]: E1206 16:17:04.488317 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:17:06 crc kubenswrapper[4813]: I1206 16:17:06.386866 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wc55n" Dec 06 16:17:06 crc kubenswrapper[4813]: I1206 16:17:06.387222 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wc55n" Dec 06 16:17:06 crc kubenswrapper[4813]: I1206 16:17:06.458548 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wc55n" Dec 06 16:17:06 crc kubenswrapper[4813]: I1206 16:17:06.560645 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wc55n" Dec 06 16:17:07 crc kubenswrapper[4813]: I1206 16:17:07.456029 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wc55n"] Dec 06 16:17:07 crc kubenswrapper[4813]: I1206 16:17:07.494924 4813 generic.go:334] "Generic (PLEG): container finished" podID="97035b28-6546-4c6b-b273-085f7ca5bb4f" containerID="1d69d25070dc9c6f430d5b7ac80b7e7769f0024ef2d7b5c0c42c31965adc25d9" exitCode=0 Dec 06 16:17:07 crc kubenswrapper[4813]: I1206 16:17:07.495063 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h7rjq" event={"ID":"97035b28-6546-4c6b-b273-085f7ca5bb4f","Type":"ContainerDied","Data":"1d69d25070dc9c6f430d5b7ac80b7e7769f0024ef2d7b5c0c42c31965adc25d9"} Dec 06 16:17:08 crc kubenswrapper[4813]: I1206 16:17:08.504443 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wc55n" podUID="1afabcb9-e54f-4951-a71c-1dc4aff14b9b" containerName="registry-server" containerID="cri-o://da84621a47fe6a1e220bf28a183562bc64d6584b538c1b7ac7f2c9bf385ad8c5" gracePeriod=2 Dec 06 16:17:08 crc kubenswrapper[4813]: E1206 16:17:08.712522 4813 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1afabcb9_e54f_4951_a71c_1dc4aff14b9b.slice/crio-conmon-da84621a47fe6a1e220bf28a183562bc64d6584b538c1b7ac7f2c9bf385ad8c5.scope\": RecentStats: unable to find data in memory cache]" Dec 06 16:17:08 crc kubenswrapper[4813]: I1206 16:17:08.957975 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wc55n" Dec 06 16:17:08 crc kubenswrapper[4813]: I1206 16:17:08.981743 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wjgn\" (UniqueName: \"kubernetes.io/projected/1afabcb9-e54f-4951-a71c-1dc4aff14b9b-kube-api-access-6wjgn\") pod \"1afabcb9-e54f-4951-a71c-1dc4aff14b9b\" (UID: \"1afabcb9-e54f-4951-a71c-1dc4aff14b9b\") " Dec 06 16:17:08 crc kubenswrapper[4813]: I1206 16:17:08.981915 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1afabcb9-e54f-4951-a71c-1dc4aff14b9b-catalog-content\") pod \"1afabcb9-e54f-4951-a71c-1dc4aff14b9b\" (UID: \"1afabcb9-e54f-4951-a71c-1dc4aff14b9b\") " Dec 06 16:17:08 crc kubenswrapper[4813]: I1206 16:17:08.981979 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1afabcb9-e54f-4951-a71c-1dc4aff14b9b-utilities\") pod \"1afabcb9-e54f-4951-a71c-1dc4aff14b9b\" (UID: \"1afabcb9-e54f-4951-a71c-1dc4aff14b9b\") " Dec 06 16:17:08 crc kubenswrapper[4813]: I1206 16:17:08.982896 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1afabcb9-e54f-4951-a71c-1dc4aff14b9b-utilities" (OuterVolumeSpecName: "utilities") pod "1afabcb9-e54f-4951-a71c-1dc4aff14b9b" (UID: "1afabcb9-e54f-4951-a71c-1dc4aff14b9b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.001826 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1afabcb9-e54f-4951-a71c-1dc4aff14b9b-kube-api-access-6wjgn" (OuterVolumeSpecName: "kube-api-access-6wjgn") pod "1afabcb9-e54f-4951-a71c-1dc4aff14b9b" (UID: "1afabcb9-e54f-4951-a71c-1dc4aff14b9b"). InnerVolumeSpecName "kube-api-access-6wjgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.054345 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1afabcb9-e54f-4951-a71c-1dc4aff14b9b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1afabcb9-e54f-4951-a71c-1dc4aff14b9b" (UID: "1afabcb9-e54f-4951-a71c-1dc4aff14b9b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.073567 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h7rjq" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.083869 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1afabcb9-e54f-4951-a71c-1dc4aff14b9b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.083898 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1afabcb9-e54f-4951-a71c-1dc4aff14b9b-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.083909 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wjgn\" (UniqueName: \"kubernetes.io/projected/1afabcb9-e54f-4951-a71c-1dc4aff14b9b-kube-api-access-6wjgn\") on node \"crc\" DevicePath \"\"" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.184614 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/97035b28-6546-4c6b-b273-085f7ca5bb4f-inventory\") pod \"97035b28-6546-4c6b-b273-085f7ca5bb4f\" (UID: \"97035b28-6546-4c6b-b273-085f7ca5bb4f\") " Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.184715 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5k4z4\" (UniqueName: \"kubernetes.io/projected/97035b28-6546-4c6b-b273-085f7ca5bb4f-kube-api-access-5k4z4\") pod \"97035b28-6546-4c6b-b273-085f7ca5bb4f\" (UID: \"97035b28-6546-4c6b-b273-085f7ca5bb4f\") " Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.184765 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/97035b28-6546-4c6b-b273-085f7ca5bb4f-ssh-key\") pod \"97035b28-6546-4c6b-b273-085f7ca5bb4f\" (UID: \"97035b28-6546-4c6b-b273-085f7ca5bb4f\") " Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.190710 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97035b28-6546-4c6b-b273-085f7ca5bb4f-kube-api-access-5k4z4" (OuterVolumeSpecName: "kube-api-access-5k4z4") pod "97035b28-6546-4c6b-b273-085f7ca5bb4f" (UID: "97035b28-6546-4c6b-b273-085f7ca5bb4f"). InnerVolumeSpecName "kube-api-access-5k4z4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.223989 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97035b28-6546-4c6b-b273-085f7ca5bb4f-inventory" (OuterVolumeSpecName: "inventory") pod "97035b28-6546-4c6b-b273-085f7ca5bb4f" (UID: "97035b28-6546-4c6b-b273-085f7ca5bb4f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.229426 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97035b28-6546-4c6b-b273-085f7ca5bb4f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "97035b28-6546-4c6b-b273-085f7ca5bb4f" (UID: "97035b28-6546-4c6b-b273-085f7ca5bb4f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.288481 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/97035b28-6546-4c6b-b273-085f7ca5bb4f-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.288511 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5k4z4\" (UniqueName: \"kubernetes.io/projected/97035b28-6546-4c6b-b273-085f7ca5bb4f-kube-api-access-5k4z4\") on node \"crc\" DevicePath \"\"" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.288522 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/97035b28-6546-4c6b-b273-085f7ca5bb4f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.513503 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h7rjq" event={"ID":"97035b28-6546-4c6b-b273-085f7ca5bb4f","Type":"ContainerDied","Data":"4994fd7220ab0e21b7b620e7fd779fac80ab8da38b047704c077bbe0e97295d4"} Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.513808 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4994fd7220ab0e21b7b620e7fd779fac80ab8da38b047704c077bbe0e97295d4" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.513552 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-h7rjq" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.515680 4813 generic.go:334] "Generic (PLEG): container finished" podID="1afabcb9-e54f-4951-a71c-1dc4aff14b9b" containerID="da84621a47fe6a1e220bf28a183562bc64d6584b538c1b7ac7f2c9bf385ad8c5" exitCode=0 Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.515706 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wc55n" event={"ID":"1afabcb9-e54f-4951-a71c-1dc4aff14b9b","Type":"ContainerDied","Data":"da84621a47fe6a1e220bf28a183562bc64d6584b538c1b7ac7f2c9bf385ad8c5"} Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.515723 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wc55n" event={"ID":"1afabcb9-e54f-4951-a71c-1dc4aff14b9b","Type":"ContainerDied","Data":"ce17840c0a664397163fc586c4c96760372e07b06cbeb9fee6555c86d40cccf0"} Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.515743 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wc55n" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.515762 4813 scope.go:117] "RemoveContainer" containerID="da84621a47fe6a1e220bf28a183562bc64d6584b538c1b7ac7f2c9bf385ad8c5" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.566026 4813 scope.go:117] "RemoveContainer" containerID="c922d922a9d7d740ce0dbd4c899b50d15093a211c05d31908e5a4f34ca7a8096" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.577048 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wc55n"] Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.597441 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wc55n"] Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.599996 4813 scope.go:117] "RemoveContainer" containerID="5c69c86a391865121ffe076785973148cd92398374e59b2a0eeaeb12216499b6" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.622453 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5"] Dec 06 16:17:09 crc kubenswrapper[4813]: E1206 16:17:09.633454 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1afabcb9-e54f-4951-a71c-1dc4aff14b9b" containerName="extract-utilities" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.633501 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="1afabcb9-e54f-4951-a71c-1dc4aff14b9b" containerName="extract-utilities" Dec 06 16:17:09 crc kubenswrapper[4813]: E1206 16:17:09.633521 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97035b28-6546-4c6b-b273-085f7ca5bb4f" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.633530 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="97035b28-6546-4c6b-b273-085f7ca5bb4f" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 06 16:17:09 crc kubenswrapper[4813]: E1206 16:17:09.633546 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1afabcb9-e54f-4951-a71c-1dc4aff14b9b" containerName="registry-server" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.633553 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="1afabcb9-e54f-4951-a71c-1dc4aff14b9b" containerName="registry-server" Dec 06 16:17:09 crc kubenswrapper[4813]: E1206 16:17:09.633571 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1afabcb9-e54f-4951-a71c-1dc4aff14b9b" containerName="extract-content" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.633580 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="1afabcb9-e54f-4951-a71c-1dc4aff14b9b" containerName="extract-content" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.634064 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="1afabcb9-e54f-4951-a71c-1dc4aff14b9b" containerName="registry-server" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.634088 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="97035b28-6546-4c6b-b273-085f7ca5bb4f" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.639003 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5"] Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.639107 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.654367 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jvkhf" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.654702 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.654876 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.655058 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.686654 4813 scope.go:117] "RemoveContainer" containerID="da84621a47fe6a1e220bf28a183562bc64d6584b538c1b7ac7f2c9bf385ad8c5" Dec 06 16:17:09 crc kubenswrapper[4813]: E1206 16:17:09.687129 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da84621a47fe6a1e220bf28a183562bc64d6584b538c1b7ac7f2c9bf385ad8c5\": container with ID starting with da84621a47fe6a1e220bf28a183562bc64d6584b538c1b7ac7f2c9bf385ad8c5 not found: ID does not exist" containerID="da84621a47fe6a1e220bf28a183562bc64d6584b538c1b7ac7f2c9bf385ad8c5" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.687167 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da84621a47fe6a1e220bf28a183562bc64d6584b538c1b7ac7f2c9bf385ad8c5"} err="failed to get container status \"da84621a47fe6a1e220bf28a183562bc64d6584b538c1b7ac7f2c9bf385ad8c5\": rpc error: code = NotFound desc = could not find container \"da84621a47fe6a1e220bf28a183562bc64d6584b538c1b7ac7f2c9bf385ad8c5\": container with ID starting with da84621a47fe6a1e220bf28a183562bc64d6584b538c1b7ac7f2c9bf385ad8c5 not found: ID does not exist" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.687193 4813 scope.go:117] "RemoveContainer" containerID="c922d922a9d7d740ce0dbd4c899b50d15093a211c05d31908e5a4f34ca7a8096" Dec 06 16:17:09 crc kubenswrapper[4813]: E1206 16:17:09.687684 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c922d922a9d7d740ce0dbd4c899b50d15093a211c05d31908e5a4f34ca7a8096\": container with ID starting with c922d922a9d7d740ce0dbd4c899b50d15093a211c05d31908e5a4f34ca7a8096 not found: ID does not exist" containerID="c922d922a9d7d740ce0dbd4c899b50d15093a211c05d31908e5a4f34ca7a8096" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.687734 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c922d922a9d7d740ce0dbd4c899b50d15093a211c05d31908e5a4f34ca7a8096"} err="failed to get container status \"c922d922a9d7d740ce0dbd4c899b50d15093a211c05d31908e5a4f34ca7a8096\": rpc error: code = NotFound desc = could not find container \"c922d922a9d7d740ce0dbd4c899b50d15093a211c05d31908e5a4f34ca7a8096\": container with ID starting with c922d922a9d7d740ce0dbd4c899b50d15093a211c05d31908e5a4f34ca7a8096 not found: ID does not exist" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.687765 4813 scope.go:117] "RemoveContainer" containerID="5c69c86a391865121ffe076785973148cd92398374e59b2a0eeaeb12216499b6" Dec 06 16:17:09 crc kubenswrapper[4813]: E1206 16:17:09.688215 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c69c86a391865121ffe076785973148cd92398374e59b2a0eeaeb12216499b6\": container with ID starting with 5c69c86a391865121ffe076785973148cd92398374e59b2a0eeaeb12216499b6 not found: ID does not exist" containerID="5c69c86a391865121ffe076785973148cd92398374e59b2a0eeaeb12216499b6" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.688270 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c69c86a391865121ffe076785973148cd92398374e59b2a0eeaeb12216499b6"} err="failed to get container status \"5c69c86a391865121ffe076785973148cd92398374e59b2a0eeaeb12216499b6\": rpc error: code = NotFound desc = could not find container \"5c69c86a391865121ffe076785973148cd92398374e59b2a0eeaeb12216499b6\": container with ID starting with 5c69c86a391865121ffe076785973148cd92398374e59b2a0eeaeb12216499b6 not found: ID does not exist" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.697427 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ddaea40-3167-4f8b-8de0-756048946c8c-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5\" (UID: \"0ddaea40-3167-4f8b-8de0-756048946c8c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.697596 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9968\" (UniqueName: \"kubernetes.io/projected/0ddaea40-3167-4f8b-8de0-756048946c8c-kube-api-access-s9968\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5\" (UID: \"0ddaea40-3167-4f8b-8de0-756048946c8c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.697797 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ddaea40-3167-4f8b-8de0-756048946c8c-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5\" (UID: \"0ddaea40-3167-4f8b-8de0-756048946c8c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.799015 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9968\" (UniqueName: \"kubernetes.io/projected/0ddaea40-3167-4f8b-8de0-756048946c8c-kube-api-access-s9968\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5\" (UID: \"0ddaea40-3167-4f8b-8de0-756048946c8c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.799109 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ddaea40-3167-4f8b-8de0-756048946c8c-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5\" (UID: \"0ddaea40-3167-4f8b-8de0-756048946c8c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.799160 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ddaea40-3167-4f8b-8de0-756048946c8c-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5\" (UID: \"0ddaea40-3167-4f8b-8de0-756048946c8c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.804063 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ddaea40-3167-4f8b-8de0-756048946c8c-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5\" (UID: \"0ddaea40-3167-4f8b-8de0-756048946c8c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.808993 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ddaea40-3167-4f8b-8de0-756048946c8c-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5\" (UID: \"0ddaea40-3167-4f8b-8de0-756048946c8c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5" Dec 06 16:17:09 crc kubenswrapper[4813]: I1206 16:17:09.817754 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9968\" (UniqueName: \"kubernetes.io/projected/0ddaea40-3167-4f8b-8de0-756048946c8c-kube-api-access-s9968\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5\" (UID: \"0ddaea40-3167-4f8b-8de0-756048946c8c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5" Dec 06 16:17:10 crc kubenswrapper[4813]: I1206 16:17:10.021209 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5" Dec 06 16:17:10 crc kubenswrapper[4813]: I1206 16:17:10.501293 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1afabcb9-e54f-4951-a71c-1dc4aff14b9b" path="/var/lib/kubelet/pods/1afabcb9-e54f-4951-a71c-1dc4aff14b9b/volumes" Dec 06 16:17:10 crc kubenswrapper[4813]: I1206 16:17:10.605587 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5"] Dec 06 16:17:11 crc kubenswrapper[4813]: I1206 16:17:11.535742 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5" event={"ID":"0ddaea40-3167-4f8b-8de0-756048946c8c","Type":"ContainerStarted","Data":"cca2b096e13360bf49ac818118fd3b5b1210f4d6161f9b55d78044a8aef86662"} Dec 06 16:17:11 crc kubenswrapper[4813]: I1206 16:17:11.536637 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5" event={"ID":"0ddaea40-3167-4f8b-8de0-756048946c8c","Type":"ContainerStarted","Data":"b9709e3752c0c9d0c94edc42c549b36fe36f4a91f2272c6a4154103b78157c94"} Dec 06 16:17:11 crc kubenswrapper[4813]: I1206 16:17:11.567204 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5" podStartSLOduration=2.087448844 podStartE2EDuration="2.567184066s" podCreationTimestamp="2025-12-06 16:17:09 +0000 UTC" firstStartedPulling="2025-12-06 16:17:10.632413621 +0000 UTC m=+1870.523293207" lastFinishedPulling="2025-12-06 16:17:11.112148813 +0000 UTC m=+1871.003028429" observedRunningTime="2025-12-06 16:17:11.552616894 +0000 UTC m=+1871.443496510" watchObservedRunningTime="2025-12-06 16:17:11.567184066 +0000 UTC m=+1871.458063652" Dec 06 16:17:18 crc kubenswrapper[4813]: I1206 16:17:18.486568 4813 scope.go:117] "RemoveContainer" containerID="dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412" Dec 06 16:17:18 crc kubenswrapper[4813]: E1206 16:17:18.487283 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:17:22 crc kubenswrapper[4813]: I1206 16:17:22.650243 4813 generic.go:334] "Generic (PLEG): container finished" podID="0ddaea40-3167-4f8b-8de0-756048946c8c" containerID="cca2b096e13360bf49ac818118fd3b5b1210f4d6161f9b55d78044a8aef86662" exitCode=0 Dec 06 16:17:22 crc kubenswrapper[4813]: I1206 16:17:22.651000 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5" event={"ID":"0ddaea40-3167-4f8b-8de0-756048946c8c","Type":"ContainerDied","Data":"cca2b096e13360bf49ac818118fd3b5b1210f4d6161f9b55d78044a8aef86662"} Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.108958 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.177153 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9968\" (UniqueName: \"kubernetes.io/projected/0ddaea40-3167-4f8b-8de0-756048946c8c-kube-api-access-s9968\") pod \"0ddaea40-3167-4f8b-8de0-756048946c8c\" (UID: \"0ddaea40-3167-4f8b-8de0-756048946c8c\") " Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.177329 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ddaea40-3167-4f8b-8de0-756048946c8c-ssh-key\") pod \"0ddaea40-3167-4f8b-8de0-756048946c8c\" (UID: \"0ddaea40-3167-4f8b-8de0-756048946c8c\") " Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.177477 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ddaea40-3167-4f8b-8de0-756048946c8c-inventory\") pod \"0ddaea40-3167-4f8b-8de0-756048946c8c\" (UID: \"0ddaea40-3167-4f8b-8de0-756048946c8c\") " Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.186197 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ddaea40-3167-4f8b-8de0-756048946c8c-kube-api-access-s9968" (OuterVolumeSpecName: "kube-api-access-s9968") pod "0ddaea40-3167-4f8b-8de0-756048946c8c" (UID: "0ddaea40-3167-4f8b-8de0-756048946c8c"). InnerVolumeSpecName "kube-api-access-s9968". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.205407 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ddaea40-3167-4f8b-8de0-756048946c8c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0ddaea40-3167-4f8b-8de0-756048946c8c" (UID: "0ddaea40-3167-4f8b-8de0-756048946c8c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.215320 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ddaea40-3167-4f8b-8de0-756048946c8c-inventory" (OuterVolumeSpecName: "inventory") pod "0ddaea40-3167-4f8b-8de0-756048946c8c" (UID: "0ddaea40-3167-4f8b-8de0-756048946c8c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.279186 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9968\" (UniqueName: \"kubernetes.io/projected/0ddaea40-3167-4f8b-8de0-756048946c8c-kube-api-access-s9968\") on node \"crc\" DevicePath \"\"" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.279232 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ddaea40-3167-4f8b-8de0-756048946c8c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.279250 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ddaea40-3167-4f8b-8de0-756048946c8c-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.681039 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5" event={"ID":"0ddaea40-3167-4f8b-8de0-756048946c8c","Type":"ContainerDied","Data":"b9709e3752c0c9d0c94edc42c549b36fe36f4a91f2272c6a4154103b78157c94"} Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.681095 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9709e3752c0c9d0c94edc42c549b36fe36f4a91f2272c6a4154103b78157c94" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.681159 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.786944 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg"] Dec 06 16:17:24 crc kubenswrapper[4813]: E1206 16:17:24.787452 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ddaea40-3167-4f8b-8de0-756048946c8c" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.787471 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ddaea40-3167-4f8b-8de0-756048946c8c" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.787706 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ddaea40-3167-4f8b-8de0-756048946c8c" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.788875 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.798011 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.798059 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.798533 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.798660 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.798741 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jvkhf" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.798824 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.798941 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.802117 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.808334 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg"] Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.892762 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8c3bde1a-68be-4aaf-9c46-f5ea14578783-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.892865 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.892900 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.892927 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.892952 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.892998 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8c3bde1a-68be-4aaf-9c46-f5ea14578783-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.893023 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.893049 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.893078 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.893167 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.893209 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8c3bde1a-68be-4aaf-9c46-f5ea14578783-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.893325 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tt9xn\" (UniqueName: \"kubernetes.io/projected/8c3bde1a-68be-4aaf-9c46-f5ea14578783-kube-api-access-tt9xn\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.893354 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8c3bde1a-68be-4aaf-9c46-f5ea14578783-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.893378 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.995656 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8c3bde1a-68be-4aaf-9c46-f5ea14578783-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.995745 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.995775 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.995802 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.995825 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.995880 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8c3bde1a-68be-4aaf-9c46-f5ea14578783-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.995903 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.995925 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.995952 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.995998 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.996040 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8c3bde1a-68be-4aaf-9c46-f5ea14578783-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.996097 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tt9xn\" (UniqueName: \"kubernetes.io/projected/8c3bde1a-68be-4aaf-9c46-f5ea14578783-kube-api-access-tt9xn\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.996123 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8c3bde1a-68be-4aaf-9c46-f5ea14578783-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:24 crc kubenswrapper[4813]: I1206 16:17:24.996146 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:25 crc kubenswrapper[4813]: I1206 16:17:25.002492 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:25 crc kubenswrapper[4813]: I1206 16:17:25.003635 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:25 crc kubenswrapper[4813]: I1206 16:17:25.004076 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8c3bde1a-68be-4aaf-9c46-f5ea14578783-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:25 crc kubenswrapper[4813]: I1206 16:17:25.004182 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:25 crc kubenswrapper[4813]: I1206 16:17:25.004670 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8c3bde1a-68be-4aaf-9c46-f5ea14578783-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:25 crc kubenswrapper[4813]: I1206 16:17:25.005206 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8c3bde1a-68be-4aaf-9c46-f5ea14578783-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:25 crc kubenswrapper[4813]: I1206 16:17:25.005730 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:25 crc kubenswrapper[4813]: I1206 16:17:25.006208 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:25 crc kubenswrapper[4813]: I1206 16:17:25.008641 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:25 crc kubenswrapper[4813]: I1206 16:17:25.011928 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:25 crc kubenswrapper[4813]: I1206 16:17:25.016840 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tt9xn\" (UniqueName: \"kubernetes.io/projected/8c3bde1a-68be-4aaf-9c46-f5ea14578783-kube-api-access-tt9xn\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:25 crc kubenswrapper[4813]: I1206 16:17:25.016941 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:25 crc kubenswrapper[4813]: I1206 16:17:25.017809 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:25 crc kubenswrapper[4813]: I1206 16:17:25.017950 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8c3bde1a-68be-4aaf-9c46-f5ea14578783-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:25 crc kubenswrapper[4813]: I1206 16:17:25.107110 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:17:25 crc kubenswrapper[4813]: I1206 16:17:25.435044 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg"] Dec 06 16:17:25 crc kubenswrapper[4813]: I1206 16:17:25.694462 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" event={"ID":"8c3bde1a-68be-4aaf-9c46-f5ea14578783","Type":"ContainerStarted","Data":"21390ed6df94346803662da5819c94324fcda11045cf4fafa11974fbb3518124"} Dec 06 16:17:26 crc kubenswrapper[4813]: I1206 16:17:26.706819 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" event={"ID":"8c3bde1a-68be-4aaf-9c46-f5ea14578783","Type":"ContainerStarted","Data":"87574ec96aa15e4c1570596b534a6fd59fad30c9832d724985701a22c040428f"} Dec 06 16:17:26 crc kubenswrapper[4813]: I1206 16:17:26.751740 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" podStartSLOduration=2.340317269 podStartE2EDuration="2.751722857s" podCreationTimestamp="2025-12-06 16:17:24 +0000 UTC" firstStartedPulling="2025-12-06 16:17:25.445465442 +0000 UTC m=+1885.336345018" lastFinishedPulling="2025-12-06 16:17:25.85687103 +0000 UTC m=+1885.747750606" observedRunningTime="2025-12-06 16:17:26.736530488 +0000 UTC m=+1886.627410104" watchObservedRunningTime="2025-12-06 16:17:26.751722857 +0000 UTC m=+1886.642602433" Dec 06 16:17:30 crc kubenswrapper[4813]: I1206 16:17:30.492983 4813 scope.go:117] "RemoveContainer" containerID="dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412" Dec 06 16:17:30 crc kubenswrapper[4813]: I1206 16:17:30.758231 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerStarted","Data":"f475b8b00603d232fa96fab484dd38d82f5207920b09491a14705e92bb65713d"} Dec 06 16:18:12 crc kubenswrapper[4813]: I1206 16:18:12.198005 4813 generic.go:334] "Generic (PLEG): container finished" podID="8c3bde1a-68be-4aaf-9c46-f5ea14578783" containerID="87574ec96aa15e4c1570596b534a6fd59fad30c9832d724985701a22c040428f" exitCode=0 Dec 06 16:18:12 crc kubenswrapper[4813]: I1206 16:18:12.198146 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" event={"ID":"8c3bde1a-68be-4aaf-9c46-f5ea14578783","Type":"ContainerDied","Data":"87574ec96aa15e4c1570596b534a6fd59fad30c9832d724985701a22c040428f"} Dec 06 16:18:13 crc kubenswrapper[4813]: I1206 16:18:13.820914 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:18:13 crc kubenswrapper[4813]: I1206 16:18:13.975526 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-neutron-metadata-combined-ca-bundle\") pod \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " Dec 06 16:18:13 crc kubenswrapper[4813]: I1206 16:18:13.975611 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8c3bde1a-68be-4aaf-9c46-f5ea14578783-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " Dec 06 16:18:13 crc kubenswrapper[4813]: I1206 16:18:13.975643 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-telemetry-combined-ca-bundle\") pod \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " Dec 06 16:18:13 crc kubenswrapper[4813]: I1206 16:18:13.975670 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8c3bde1a-68be-4aaf-9c46-f5ea14578783-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " Dec 06 16:18:13 crc kubenswrapper[4813]: I1206 16:18:13.975765 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8c3bde1a-68be-4aaf-9c46-f5ea14578783-openstack-edpm-ipam-ovn-default-certs-0\") pod \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " Dec 06 16:18:13 crc kubenswrapper[4813]: I1206 16:18:13.975873 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-ssh-key\") pod \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " Dec 06 16:18:13 crc kubenswrapper[4813]: I1206 16:18:13.975925 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8c3bde1a-68be-4aaf-9c46-f5ea14578783-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " Dec 06 16:18:13 crc kubenswrapper[4813]: I1206 16:18:13.975969 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-inventory\") pod \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " Dec 06 16:18:13 crc kubenswrapper[4813]: I1206 16:18:13.975995 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-libvirt-combined-ca-bundle\") pod \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " Dec 06 16:18:13 crc kubenswrapper[4813]: I1206 16:18:13.976037 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tt9xn\" (UniqueName: \"kubernetes.io/projected/8c3bde1a-68be-4aaf-9c46-f5ea14578783-kube-api-access-tt9xn\") pod \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " Dec 06 16:18:13 crc kubenswrapper[4813]: I1206 16:18:13.976089 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-repo-setup-combined-ca-bundle\") pod \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " Dec 06 16:18:13 crc kubenswrapper[4813]: I1206 16:18:13.976106 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-ovn-combined-ca-bundle\") pod \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " Dec 06 16:18:13 crc kubenswrapper[4813]: I1206 16:18:13.976144 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-bootstrap-combined-ca-bundle\") pod \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " Dec 06 16:18:13 crc kubenswrapper[4813]: I1206 16:18:13.976175 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-nova-combined-ca-bundle\") pod \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\" (UID: \"8c3bde1a-68be-4aaf-9c46-f5ea14578783\") " Dec 06 16:18:13 crc kubenswrapper[4813]: I1206 16:18:13.983042 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "8c3bde1a-68be-4aaf-9c46-f5ea14578783" (UID: "8c3bde1a-68be-4aaf-9c46-f5ea14578783"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:18:13 crc kubenswrapper[4813]: I1206 16:18:13.983067 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "8c3bde1a-68be-4aaf-9c46-f5ea14578783" (UID: "8c3bde1a-68be-4aaf-9c46-f5ea14578783"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:18:13 crc kubenswrapper[4813]: I1206 16:18:13.983913 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c3bde1a-68be-4aaf-9c46-f5ea14578783-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "8c3bde1a-68be-4aaf-9c46-f5ea14578783" (UID: "8c3bde1a-68be-4aaf-9c46-f5ea14578783"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:18:13 crc kubenswrapper[4813]: I1206 16:18:13.983999 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "8c3bde1a-68be-4aaf-9c46-f5ea14578783" (UID: "8c3bde1a-68be-4aaf-9c46-f5ea14578783"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:18:13 crc kubenswrapper[4813]: I1206 16:18:13.984655 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c3bde1a-68be-4aaf-9c46-f5ea14578783-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "8c3bde1a-68be-4aaf-9c46-f5ea14578783" (UID: "8c3bde1a-68be-4aaf-9c46-f5ea14578783"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:18:13 crc kubenswrapper[4813]: I1206 16:18:13.985759 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c3bde1a-68be-4aaf-9c46-f5ea14578783-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "8c3bde1a-68be-4aaf-9c46-f5ea14578783" (UID: "8c3bde1a-68be-4aaf-9c46-f5ea14578783"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:18:13 crc kubenswrapper[4813]: I1206 16:18:13.985804 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "8c3bde1a-68be-4aaf-9c46-f5ea14578783" (UID: "8c3bde1a-68be-4aaf-9c46-f5ea14578783"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:18:13 crc kubenswrapper[4813]: I1206 16:18:13.986603 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "8c3bde1a-68be-4aaf-9c46-f5ea14578783" (UID: "8c3bde1a-68be-4aaf-9c46-f5ea14578783"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:18:13 crc kubenswrapper[4813]: I1206 16:18:13.986616 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c3bde1a-68be-4aaf-9c46-f5ea14578783-kube-api-access-tt9xn" (OuterVolumeSpecName: "kube-api-access-tt9xn") pod "8c3bde1a-68be-4aaf-9c46-f5ea14578783" (UID: "8c3bde1a-68be-4aaf-9c46-f5ea14578783"). InnerVolumeSpecName "kube-api-access-tt9xn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:18:13 crc kubenswrapper[4813]: I1206 16:18:13.987162 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c3bde1a-68be-4aaf-9c46-f5ea14578783-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "8c3bde1a-68be-4aaf-9c46-f5ea14578783" (UID: "8c3bde1a-68be-4aaf-9c46-f5ea14578783"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:18:13 crc kubenswrapper[4813]: I1206 16:18:13.987801 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "8c3bde1a-68be-4aaf-9c46-f5ea14578783" (UID: "8c3bde1a-68be-4aaf-9c46-f5ea14578783"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:18:13 crc kubenswrapper[4813]: I1206 16:18:13.991981 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "8c3bde1a-68be-4aaf-9c46-f5ea14578783" (UID: "8c3bde1a-68be-4aaf-9c46-f5ea14578783"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.010133 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-inventory" (OuterVolumeSpecName: "inventory") pod "8c3bde1a-68be-4aaf-9c46-f5ea14578783" (UID: "8c3bde1a-68be-4aaf-9c46-f5ea14578783"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.016582 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8c3bde1a-68be-4aaf-9c46-f5ea14578783" (UID: "8c3bde1a-68be-4aaf-9c46-f5ea14578783"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.077732 4813 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8c3bde1a-68be-4aaf-9c46-f5ea14578783-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.077766 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.077776 4813 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.077788 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tt9xn\" (UniqueName: \"kubernetes.io/projected/8c3bde1a-68be-4aaf-9c46-f5ea14578783-kube-api-access-tt9xn\") on node \"crc\" DevicePath \"\"" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.077796 4813 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.077806 4813 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.077815 4813 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.077826 4813 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.077834 4813 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.077843 4813 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8c3bde1a-68be-4aaf-9c46-f5ea14578783-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.077852 4813 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.077862 4813 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8c3bde1a-68be-4aaf-9c46-f5ea14578783-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.077873 4813 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8c3bde1a-68be-4aaf-9c46-f5ea14578783-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.077881 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8c3bde1a-68be-4aaf-9c46-f5ea14578783-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.222234 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" event={"ID":"8c3bde1a-68be-4aaf-9c46-f5ea14578783","Type":"ContainerDied","Data":"21390ed6df94346803662da5819c94324fcda11045cf4fafa11974fbb3518124"} Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.222558 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21390ed6df94346803662da5819c94324fcda11045cf4fafa11974fbb3518124" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.222305 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.360473 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-6gtnd"] Dec 06 16:18:14 crc kubenswrapper[4813]: E1206 16:18:14.360902 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c3bde1a-68be-4aaf-9c46-f5ea14578783" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.360920 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c3bde1a-68be-4aaf-9c46-f5ea14578783" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.361115 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c3bde1a-68be-4aaf-9c46-f5ea14578783" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.361799 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6gtnd" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.366363 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.366372 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.366971 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.367503 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jvkhf" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.367560 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.370574 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-6gtnd"] Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.388869 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/ef88d823-7e10-4eee-a6b1-66c680355906-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6gtnd\" (UID: \"ef88d823-7e10-4eee-a6b1-66c680355906\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6gtnd" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.388918 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef88d823-7e10-4eee-a6b1-66c680355906-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6gtnd\" (UID: \"ef88d823-7e10-4eee-a6b1-66c680355906\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6gtnd" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.389001 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7gzt\" (UniqueName: \"kubernetes.io/projected/ef88d823-7e10-4eee-a6b1-66c680355906-kube-api-access-f7gzt\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6gtnd\" (UID: \"ef88d823-7e10-4eee-a6b1-66c680355906\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6gtnd" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.389047 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef88d823-7e10-4eee-a6b1-66c680355906-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6gtnd\" (UID: \"ef88d823-7e10-4eee-a6b1-66c680355906\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6gtnd" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.389088 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef88d823-7e10-4eee-a6b1-66c680355906-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6gtnd\" (UID: \"ef88d823-7e10-4eee-a6b1-66c680355906\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6gtnd" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.490170 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7gzt\" (UniqueName: \"kubernetes.io/projected/ef88d823-7e10-4eee-a6b1-66c680355906-kube-api-access-f7gzt\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6gtnd\" (UID: \"ef88d823-7e10-4eee-a6b1-66c680355906\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6gtnd" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.490772 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef88d823-7e10-4eee-a6b1-66c680355906-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6gtnd\" (UID: \"ef88d823-7e10-4eee-a6b1-66c680355906\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6gtnd" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.491478 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef88d823-7e10-4eee-a6b1-66c680355906-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6gtnd\" (UID: \"ef88d823-7e10-4eee-a6b1-66c680355906\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6gtnd" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.492196 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef88d823-7e10-4eee-a6b1-66c680355906-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6gtnd\" (UID: \"ef88d823-7e10-4eee-a6b1-66c680355906\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6gtnd" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.492307 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/ef88d823-7e10-4eee-a6b1-66c680355906-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6gtnd\" (UID: \"ef88d823-7e10-4eee-a6b1-66c680355906\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6gtnd" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.493478 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/ef88d823-7e10-4eee-a6b1-66c680355906-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6gtnd\" (UID: \"ef88d823-7e10-4eee-a6b1-66c680355906\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6gtnd" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.496704 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef88d823-7e10-4eee-a6b1-66c680355906-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6gtnd\" (UID: \"ef88d823-7e10-4eee-a6b1-66c680355906\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6gtnd" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.516156 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7gzt\" (UniqueName: \"kubernetes.io/projected/ef88d823-7e10-4eee-a6b1-66c680355906-kube-api-access-f7gzt\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6gtnd\" (UID: \"ef88d823-7e10-4eee-a6b1-66c680355906\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6gtnd" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.516365 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef88d823-7e10-4eee-a6b1-66c680355906-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6gtnd\" (UID: \"ef88d823-7e10-4eee-a6b1-66c680355906\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6gtnd" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.516585 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef88d823-7e10-4eee-a6b1-66c680355906-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6gtnd\" (UID: \"ef88d823-7e10-4eee-a6b1-66c680355906\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6gtnd" Dec 06 16:18:14 crc kubenswrapper[4813]: I1206 16:18:14.685687 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6gtnd" Dec 06 16:18:15 crc kubenswrapper[4813]: I1206 16:18:15.314853 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-6gtnd"] Dec 06 16:18:16 crc kubenswrapper[4813]: I1206 16:18:16.249430 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6gtnd" event={"ID":"ef88d823-7e10-4eee-a6b1-66c680355906","Type":"ContainerStarted","Data":"c67783331a0f590dae7a36d3659b65540573a2e1dc48696dfa0ddc64dedd9ab9"} Dec 06 16:18:16 crc kubenswrapper[4813]: I1206 16:18:16.250410 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6gtnd" event={"ID":"ef88d823-7e10-4eee-a6b1-66c680355906","Type":"ContainerStarted","Data":"c033e7bebcb871b978be8c06c06f5842b5e43811ebd1ec676def1a4de5fc8852"} Dec 06 16:18:16 crc kubenswrapper[4813]: I1206 16:18:16.280160 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6gtnd" podStartSLOduration=1.77203142 podStartE2EDuration="2.280145086s" podCreationTimestamp="2025-12-06 16:18:14 +0000 UTC" firstStartedPulling="2025-12-06 16:18:15.330612514 +0000 UTC m=+1935.221492090" lastFinishedPulling="2025-12-06 16:18:15.83872618 +0000 UTC m=+1935.729605756" observedRunningTime="2025-12-06 16:18:16.276675075 +0000 UTC m=+1936.167554651" watchObservedRunningTime="2025-12-06 16:18:16.280145086 +0000 UTC m=+1936.171024652" Dec 06 16:19:31 crc kubenswrapper[4813]: I1206 16:19:31.057599 4813 generic.go:334] "Generic (PLEG): container finished" podID="ef88d823-7e10-4eee-a6b1-66c680355906" containerID="c67783331a0f590dae7a36d3659b65540573a2e1dc48696dfa0ddc64dedd9ab9" exitCode=0 Dec 06 16:19:31 crc kubenswrapper[4813]: I1206 16:19:31.057680 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6gtnd" event={"ID":"ef88d823-7e10-4eee-a6b1-66c680355906","Type":"ContainerDied","Data":"c67783331a0f590dae7a36d3659b65540573a2e1dc48696dfa0ddc64dedd9ab9"} Dec 06 16:19:32 crc kubenswrapper[4813]: I1206 16:19:32.503061 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6gtnd" Dec 06 16:19:32 crc kubenswrapper[4813]: I1206 16:19:32.618411 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/ef88d823-7e10-4eee-a6b1-66c680355906-ovncontroller-config-0\") pod \"ef88d823-7e10-4eee-a6b1-66c680355906\" (UID: \"ef88d823-7e10-4eee-a6b1-66c680355906\") " Dec 06 16:19:32 crc kubenswrapper[4813]: I1206 16:19:32.618547 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef88d823-7e10-4eee-a6b1-66c680355906-ovn-combined-ca-bundle\") pod \"ef88d823-7e10-4eee-a6b1-66c680355906\" (UID: \"ef88d823-7e10-4eee-a6b1-66c680355906\") " Dec 06 16:19:32 crc kubenswrapper[4813]: I1206 16:19:32.618583 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef88d823-7e10-4eee-a6b1-66c680355906-ssh-key\") pod \"ef88d823-7e10-4eee-a6b1-66c680355906\" (UID: \"ef88d823-7e10-4eee-a6b1-66c680355906\") " Dec 06 16:19:32 crc kubenswrapper[4813]: I1206 16:19:32.618666 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef88d823-7e10-4eee-a6b1-66c680355906-inventory\") pod \"ef88d823-7e10-4eee-a6b1-66c680355906\" (UID: \"ef88d823-7e10-4eee-a6b1-66c680355906\") " Dec 06 16:19:32 crc kubenswrapper[4813]: I1206 16:19:32.618768 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7gzt\" (UniqueName: \"kubernetes.io/projected/ef88d823-7e10-4eee-a6b1-66c680355906-kube-api-access-f7gzt\") pod \"ef88d823-7e10-4eee-a6b1-66c680355906\" (UID: \"ef88d823-7e10-4eee-a6b1-66c680355906\") " Dec 06 16:19:32 crc kubenswrapper[4813]: I1206 16:19:32.626886 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef88d823-7e10-4eee-a6b1-66c680355906-kube-api-access-f7gzt" (OuterVolumeSpecName: "kube-api-access-f7gzt") pod "ef88d823-7e10-4eee-a6b1-66c680355906" (UID: "ef88d823-7e10-4eee-a6b1-66c680355906"). InnerVolumeSpecName "kube-api-access-f7gzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:19:32 crc kubenswrapper[4813]: I1206 16:19:32.632416 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef88d823-7e10-4eee-a6b1-66c680355906-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "ef88d823-7e10-4eee-a6b1-66c680355906" (UID: "ef88d823-7e10-4eee-a6b1-66c680355906"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:19:32 crc kubenswrapper[4813]: I1206 16:19:32.646253 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef88d823-7e10-4eee-a6b1-66c680355906-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "ef88d823-7e10-4eee-a6b1-66c680355906" (UID: "ef88d823-7e10-4eee-a6b1-66c680355906"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:19:32 crc kubenswrapper[4813]: I1206 16:19:32.646833 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef88d823-7e10-4eee-a6b1-66c680355906-inventory" (OuterVolumeSpecName: "inventory") pod "ef88d823-7e10-4eee-a6b1-66c680355906" (UID: "ef88d823-7e10-4eee-a6b1-66c680355906"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:19:32 crc kubenswrapper[4813]: I1206 16:19:32.668068 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef88d823-7e10-4eee-a6b1-66c680355906-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ef88d823-7e10-4eee-a6b1-66c680355906" (UID: "ef88d823-7e10-4eee-a6b1-66c680355906"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:19:32 crc kubenswrapper[4813]: I1206 16:19:32.721308 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7gzt\" (UniqueName: \"kubernetes.io/projected/ef88d823-7e10-4eee-a6b1-66c680355906-kube-api-access-f7gzt\") on node \"crc\" DevicePath \"\"" Dec 06 16:19:32 crc kubenswrapper[4813]: I1206 16:19:32.721338 4813 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/ef88d823-7e10-4eee-a6b1-66c680355906-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 16:19:32 crc kubenswrapper[4813]: I1206 16:19:32.721349 4813 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef88d823-7e10-4eee-a6b1-66c680355906-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:19:32 crc kubenswrapper[4813]: I1206 16:19:32.721358 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef88d823-7e10-4eee-a6b1-66c680355906-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 16:19:32 crc kubenswrapper[4813]: I1206 16:19:32.721367 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef88d823-7e10-4eee-a6b1-66c680355906-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.078391 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6gtnd" event={"ID":"ef88d823-7e10-4eee-a6b1-66c680355906","Type":"ContainerDied","Data":"c033e7bebcb871b978be8c06c06f5842b5e43811ebd1ec676def1a4de5fc8852"} Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.078451 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c033e7bebcb871b978be8c06c06f5842b5e43811ebd1ec676def1a4de5fc8852" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.078533 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6gtnd" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.203969 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk"] Dec 06 16:19:33 crc kubenswrapper[4813]: E1206 16:19:33.204332 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef88d823-7e10-4eee-a6b1-66c680355906" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.204349 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef88d823-7e10-4eee-a6b1-66c680355906" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.204567 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef88d823-7e10-4eee-a6b1-66c680355906" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.205311 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.208903 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.209867 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.210631 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jvkhf" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.211105 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.225394 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk"] Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.225920 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.226145 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.330760 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pncm5\" (UniqueName: \"kubernetes.io/projected/a11c7f97-21d3-48b1-9640-529df9d946b8-kube-api-access-pncm5\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk\" (UID: \"a11c7f97-21d3-48b1-9640-529df9d946b8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.330840 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a11c7f97-21d3-48b1-9640-529df9d946b8-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk\" (UID: \"a11c7f97-21d3-48b1-9640-529df9d946b8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.331821 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a11c7f97-21d3-48b1-9640-529df9d946b8-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk\" (UID: \"a11c7f97-21d3-48b1-9640-529df9d946b8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.331937 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a11c7f97-21d3-48b1-9640-529df9d946b8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk\" (UID: \"a11c7f97-21d3-48b1-9640-529df9d946b8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.332002 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a11c7f97-21d3-48b1-9640-529df9d946b8-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk\" (UID: \"a11c7f97-21d3-48b1-9640-529df9d946b8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.332149 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a11c7f97-21d3-48b1-9640-529df9d946b8-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk\" (UID: \"a11c7f97-21d3-48b1-9640-529df9d946b8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.434489 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a11c7f97-21d3-48b1-9640-529df9d946b8-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk\" (UID: \"a11c7f97-21d3-48b1-9640-529df9d946b8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.434871 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a11c7f97-21d3-48b1-9640-529df9d946b8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk\" (UID: \"a11c7f97-21d3-48b1-9640-529df9d946b8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.434928 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a11c7f97-21d3-48b1-9640-529df9d946b8-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk\" (UID: \"a11c7f97-21d3-48b1-9640-529df9d946b8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.435038 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a11c7f97-21d3-48b1-9640-529df9d946b8-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk\" (UID: \"a11c7f97-21d3-48b1-9640-529df9d946b8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.435099 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pncm5\" (UniqueName: \"kubernetes.io/projected/a11c7f97-21d3-48b1-9640-529df9d946b8-kube-api-access-pncm5\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk\" (UID: \"a11c7f97-21d3-48b1-9640-529df9d946b8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.435210 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a11c7f97-21d3-48b1-9640-529df9d946b8-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk\" (UID: \"a11c7f97-21d3-48b1-9640-529df9d946b8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.446198 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a11c7f97-21d3-48b1-9640-529df9d946b8-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk\" (UID: \"a11c7f97-21d3-48b1-9640-529df9d946b8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.448838 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a11c7f97-21d3-48b1-9640-529df9d946b8-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk\" (UID: \"a11c7f97-21d3-48b1-9640-529df9d946b8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.449797 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a11c7f97-21d3-48b1-9640-529df9d946b8-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk\" (UID: \"a11c7f97-21d3-48b1-9640-529df9d946b8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.449991 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a11c7f97-21d3-48b1-9640-529df9d946b8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk\" (UID: \"a11c7f97-21d3-48b1-9640-529df9d946b8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.463727 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a11c7f97-21d3-48b1-9640-529df9d946b8-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk\" (UID: \"a11c7f97-21d3-48b1-9640-529df9d946b8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.467950 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pncm5\" (UniqueName: \"kubernetes.io/projected/a11c7f97-21d3-48b1-9640-529df9d946b8-kube-api-access-pncm5\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk\" (UID: \"a11c7f97-21d3-48b1-9640-529df9d946b8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk" Dec 06 16:19:33 crc kubenswrapper[4813]: I1206 16:19:33.522899 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk" Dec 06 16:19:34 crc kubenswrapper[4813]: I1206 16:19:34.154045 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk"] Dec 06 16:19:34 crc kubenswrapper[4813]: W1206 16:19:34.159869 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda11c7f97_21d3_48b1_9640_529df9d946b8.slice/crio-3b868fc653196d9902cbc730841def5ce32e511a10537fb08762d0f39e60b79c WatchSource:0}: Error finding container 3b868fc653196d9902cbc730841def5ce32e511a10537fb08762d0f39e60b79c: Status 404 returned error can't find the container with id 3b868fc653196d9902cbc730841def5ce32e511a10537fb08762d0f39e60b79c Dec 06 16:19:35 crc kubenswrapper[4813]: I1206 16:19:35.101137 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk" event={"ID":"a11c7f97-21d3-48b1-9640-529df9d946b8","Type":"ContainerStarted","Data":"f3d123ee8868dbe9baeb8cddfc96d4f75245e58b4cf4ba9390bdb5593800211c"} Dec 06 16:19:35 crc kubenswrapper[4813]: I1206 16:19:35.101485 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk" event={"ID":"a11c7f97-21d3-48b1-9640-529df9d946b8","Type":"ContainerStarted","Data":"3b868fc653196d9902cbc730841def5ce32e511a10537fb08762d0f39e60b79c"} Dec 06 16:19:35 crc kubenswrapper[4813]: I1206 16:19:35.130535 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk" podStartSLOduration=1.6646311580000002 podStartE2EDuration="2.130515422s" podCreationTimestamp="2025-12-06 16:19:33 +0000 UTC" firstStartedPulling="2025-12-06 16:19:34.163096912 +0000 UTC m=+2014.053976508" lastFinishedPulling="2025-12-06 16:19:34.628981196 +0000 UTC m=+2014.519860772" observedRunningTime="2025-12-06 16:19:35.118377255 +0000 UTC m=+2015.009256821" watchObservedRunningTime="2025-12-06 16:19:35.130515422 +0000 UTC m=+2015.021395008" Dec 06 16:19:49 crc kubenswrapper[4813]: I1206 16:19:49.428030 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:19:49 crc kubenswrapper[4813]: I1206 16:19:49.428472 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:20:19 crc kubenswrapper[4813]: I1206 16:20:19.428451 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:20:19 crc kubenswrapper[4813]: I1206 16:20:19.429156 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:20:31 crc kubenswrapper[4813]: I1206 16:20:31.696617 4813 generic.go:334] "Generic (PLEG): container finished" podID="a11c7f97-21d3-48b1-9640-529df9d946b8" containerID="f3d123ee8868dbe9baeb8cddfc96d4f75245e58b4cf4ba9390bdb5593800211c" exitCode=0 Dec 06 16:20:31 crc kubenswrapper[4813]: I1206 16:20:31.696715 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk" event={"ID":"a11c7f97-21d3-48b1-9640-529df9d946b8","Type":"ContainerDied","Data":"f3d123ee8868dbe9baeb8cddfc96d4f75245e58b4cf4ba9390bdb5593800211c"} Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.219628 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk" Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.290152 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pncm5\" (UniqueName: \"kubernetes.io/projected/a11c7f97-21d3-48b1-9640-529df9d946b8-kube-api-access-pncm5\") pod \"a11c7f97-21d3-48b1-9640-529df9d946b8\" (UID: \"a11c7f97-21d3-48b1-9640-529df9d946b8\") " Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.290195 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a11c7f97-21d3-48b1-9640-529df9d946b8-neutron-metadata-combined-ca-bundle\") pod \"a11c7f97-21d3-48b1-9640-529df9d946b8\" (UID: \"a11c7f97-21d3-48b1-9640-529df9d946b8\") " Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.290237 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a11c7f97-21d3-48b1-9640-529df9d946b8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"a11c7f97-21d3-48b1-9640-529df9d946b8\" (UID: \"a11c7f97-21d3-48b1-9640-529df9d946b8\") " Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.290304 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a11c7f97-21d3-48b1-9640-529df9d946b8-inventory\") pod \"a11c7f97-21d3-48b1-9640-529df9d946b8\" (UID: \"a11c7f97-21d3-48b1-9640-529df9d946b8\") " Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.290366 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a11c7f97-21d3-48b1-9640-529df9d946b8-nova-metadata-neutron-config-0\") pod \"a11c7f97-21d3-48b1-9640-529df9d946b8\" (UID: \"a11c7f97-21d3-48b1-9640-529df9d946b8\") " Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.290390 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a11c7f97-21d3-48b1-9640-529df9d946b8-ssh-key\") pod \"a11c7f97-21d3-48b1-9640-529df9d946b8\" (UID: \"a11c7f97-21d3-48b1-9640-529df9d946b8\") " Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.299598 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a11c7f97-21d3-48b1-9640-529df9d946b8-kube-api-access-pncm5" (OuterVolumeSpecName: "kube-api-access-pncm5") pod "a11c7f97-21d3-48b1-9640-529df9d946b8" (UID: "a11c7f97-21d3-48b1-9640-529df9d946b8"). InnerVolumeSpecName "kube-api-access-pncm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.310585 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a11c7f97-21d3-48b1-9640-529df9d946b8-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "a11c7f97-21d3-48b1-9640-529df9d946b8" (UID: "a11c7f97-21d3-48b1-9640-529df9d946b8"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.325725 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a11c7f97-21d3-48b1-9640-529df9d946b8-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "a11c7f97-21d3-48b1-9640-529df9d946b8" (UID: "a11c7f97-21d3-48b1-9640-529df9d946b8"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.331834 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a11c7f97-21d3-48b1-9640-529df9d946b8-inventory" (OuterVolumeSpecName: "inventory") pod "a11c7f97-21d3-48b1-9640-529df9d946b8" (UID: "a11c7f97-21d3-48b1-9640-529df9d946b8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.333279 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a11c7f97-21d3-48b1-9640-529df9d946b8-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "a11c7f97-21d3-48b1-9640-529df9d946b8" (UID: "a11c7f97-21d3-48b1-9640-529df9d946b8"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.361720 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a11c7f97-21d3-48b1-9640-529df9d946b8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a11c7f97-21d3-48b1-9640-529df9d946b8" (UID: "a11c7f97-21d3-48b1-9640-529df9d946b8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.392527 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pncm5\" (UniqueName: \"kubernetes.io/projected/a11c7f97-21d3-48b1-9640-529df9d946b8-kube-api-access-pncm5\") on node \"crc\" DevicePath \"\"" Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.392555 4813 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a11c7f97-21d3-48b1-9640-529df9d946b8-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.392565 4813 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a11c7f97-21d3-48b1-9640-529df9d946b8-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.392578 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a11c7f97-21d3-48b1-9640-529df9d946b8-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.392588 4813 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a11c7f97-21d3-48b1-9640-529df9d946b8-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.392597 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a11c7f97-21d3-48b1-9640-529df9d946b8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.725842 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk" event={"ID":"a11c7f97-21d3-48b1-9640-529df9d946b8","Type":"ContainerDied","Data":"3b868fc653196d9902cbc730841def5ce32e511a10537fb08762d0f39e60b79c"} Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.725900 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b868fc653196d9902cbc730841def5ce32e511a10537fb08762d0f39e60b79c" Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.725932 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk" Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.853956 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j"] Dec 06 16:20:33 crc kubenswrapper[4813]: E1206 16:20:33.854601 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a11c7f97-21d3-48b1-9640-529df9d946b8" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.854616 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a11c7f97-21d3-48b1-9640-529df9d946b8" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.854820 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a11c7f97-21d3-48b1-9640-529df9d946b8" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.855438 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j" Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.874110 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.874209 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.874429 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.874490 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.874608 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jvkhf" Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.883993 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j"] Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.905671 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bfa11078-d153-43a2-abc4-b70f491dc4a6-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j\" (UID: \"bfa11078-d153-43a2-abc4-b70f491dc4a6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j" Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.905723 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d628l\" (UniqueName: \"kubernetes.io/projected/bfa11078-d153-43a2-abc4-b70f491dc4a6-kube-api-access-d628l\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j\" (UID: \"bfa11078-d153-43a2-abc4-b70f491dc4a6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j" Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.905795 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/bfa11078-d153-43a2-abc4-b70f491dc4a6-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j\" (UID: \"bfa11078-d153-43a2-abc4-b70f491dc4a6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j" Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.905849 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bfa11078-d153-43a2-abc4-b70f491dc4a6-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j\" (UID: \"bfa11078-d153-43a2-abc4-b70f491dc4a6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j" Dec 06 16:20:33 crc kubenswrapper[4813]: I1206 16:20:33.905926 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfa11078-d153-43a2-abc4-b70f491dc4a6-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j\" (UID: \"bfa11078-d153-43a2-abc4-b70f491dc4a6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j" Dec 06 16:20:34 crc kubenswrapper[4813]: I1206 16:20:34.007099 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bfa11078-d153-43a2-abc4-b70f491dc4a6-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j\" (UID: \"bfa11078-d153-43a2-abc4-b70f491dc4a6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j" Dec 06 16:20:34 crc kubenswrapper[4813]: I1206 16:20:34.007444 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfa11078-d153-43a2-abc4-b70f491dc4a6-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j\" (UID: \"bfa11078-d153-43a2-abc4-b70f491dc4a6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j" Dec 06 16:20:34 crc kubenswrapper[4813]: I1206 16:20:34.007581 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bfa11078-d153-43a2-abc4-b70f491dc4a6-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j\" (UID: \"bfa11078-d153-43a2-abc4-b70f491dc4a6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j" Dec 06 16:20:34 crc kubenswrapper[4813]: I1206 16:20:34.007765 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d628l\" (UniqueName: \"kubernetes.io/projected/bfa11078-d153-43a2-abc4-b70f491dc4a6-kube-api-access-d628l\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j\" (UID: \"bfa11078-d153-43a2-abc4-b70f491dc4a6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j" Dec 06 16:20:34 crc kubenswrapper[4813]: I1206 16:20:34.008366 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/bfa11078-d153-43a2-abc4-b70f491dc4a6-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j\" (UID: \"bfa11078-d153-43a2-abc4-b70f491dc4a6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j" Dec 06 16:20:34 crc kubenswrapper[4813]: I1206 16:20:34.011506 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bfa11078-d153-43a2-abc4-b70f491dc4a6-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j\" (UID: \"bfa11078-d153-43a2-abc4-b70f491dc4a6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j" Dec 06 16:20:34 crc kubenswrapper[4813]: I1206 16:20:34.011817 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfa11078-d153-43a2-abc4-b70f491dc4a6-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j\" (UID: \"bfa11078-d153-43a2-abc4-b70f491dc4a6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j" Dec 06 16:20:34 crc kubenswrapper[4813]: I1206 16:20:34.011837 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bfa11078-d153-43a2-abc4-b70f491dc4a6-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j\" (UID: \"bfa11078-d153-43a2-abc4-b70f491dc4a6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j" Dec 06 16:20:34 crc kubenswrapper[4813]: I1206 16:20:34.013871 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/bfa11078-d153-43a2-abc4-b70f491dc4a6-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j\" (UID: \"bfa11078-d153-43a2-abc4-b70f491dc4a6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j" Dec 06 16:20:34 crc kubenswrapper[4813]: I1206 16:20:34.028438 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d628l\" (UniqueName: \"kubernetes.io/projected/bfa11078-d153-43a2-abc4-b70f491dc4a6-kube-api-access-d628l\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j\" (UID: \"bfa11078-d153-43a2-abc4-b70f491dc4a6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j" Dec 06 16:20:34 crc kubenswrapper[4813]: I1206 16:20:34.174910 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j" Dec 06 16:20:34 crc kubenswrapper[4813]: I1206 16:20:34.779742 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j"] Dec 06 16:20:35 crc kubenswrapper[4813]: I1206 16:20:35.749487 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j" event={"ID":"bfa11078-d153-43a2-abc4-b70f491dc4a6","Type":"ContainerStarted","Data":"814ae62233afc9678a4c073b9a4fec1cf6f15e78b47d1bfa35830574e9f5a35b"} Dec 06 16:20:35 crc kubenswrapper[4813]: I1206 16:20:35.749866 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j" event={"ID":"bfa11078-d153-43a2-abc4-b70f491dc4a6","Type":"ContainerStarted","Data":"d5914d41175b97c87ad4c0cedfe6473ab888f5a698b85a8b551c385d77dd5efd"} Dec 06 16:20:35 crc kubenswrapper[4813]: I1206 16:20:35.781151 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j" podStartSLOduration=2.275824041 podStartE2EDuration="2.781134005s" podCreationTimestamp="2025-12-06 16:20:33 +0000 UTC" firstStartedPulling="2025-12-06 16:20:34.774990953 +0000 UTC m=+2074.665870539" lastFinishedPulling="2025-12-06 16:20:35.280300897 +0000 UTC m=+2075.171180503" observedRunningTime="2025-12-06 16:20:35.773337941 +0000 UTC m=+2075.664217557" watchObservedRunningTime="2025-12-06 16:20:35.781134005 +0000 UTC m=+2075.672013581" Dec 06 16:20:49 crc kubenswrapper[4813]: I1206 16:20:49.427889 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:20:49 crc kubenswrapper[4813]: I1206 16:20:49.428540 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:20:49 crc kubenswrapper[4813]: I1206 16:20:49.428603 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" Dec 06 16:20:49 crc kubenswrapper[4813]: I1206 16:20:49.429412 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f475b8b00603d232fa96fab484dd38d82f5207920b09491a14705e92bb65713d"} pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 16:20:49 crc kubenswrapper[4813]: I1206 16:20:49.429507 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" containerID="cri-o://f475b8b00603d232fa96fab484dd38d82f5207920b09491a14705e92bb65713d" gracePeriod=600 Dec 06 16:20:49 crc kubenswrapper[4813]: I1206 16:20:49.914983 4813 generic.go:334] "Generic (PLEG): container finished" podID="d88e8bae-c055-4c55-b548-f621ff96de06" containerID="f475b8b00603d232fa96fab484dd38d82f5207920b09491a14705e92bb65713d" exitCode=0 Dec 06 16:20:49 crc kubenswrapper[4813]: I1206 16:20:49.915053 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerDied","Data":"f475b8b00603d232fa96fab484dd38d82f5207920b09491a14705e92bb65713d"} Dec 06 16:20:49 crc kubenswrapper[4813]: I1206 16:20:49.915655 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerStarted","Data":"1b8cc98443ea491aa620632d54d48664f2e8cd65b0790f1013b24051a06625ca"} Dec 06 16:20:49 crc kubenswrapper[4813]: I1206 16:20:49.915769 4813 scope.go:117] "RemoveContainer" containerID="dfeab1ba2a4820035fc4b2da49ae95526cdf6b4fac4e848436f52f5f1fd71412" Dec 06 16:21:06 crc kubenswrapper[4813]: I1206 16:21:06.631768 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7hx5h"] Dec 06 16:21:06 crc kubenswrapper[4813]: I1206 16:21:06.634430 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7hx5h" Dec 06 16:21:06 crc kubenswrapper[4813]: I1206 16:21:06.651827 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7hx5h"] Dec 06 16:21:06 crc kubenswrapper[4813]: I1206 16:21:06.697275 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6724f83-f171-4e2a-a03f-4109fc38c74a-utilities\") pod \"redhat-marketplace-7hx5h\" (UID: \"c6724f83-f171-4e2a-a03f-4109fc38c74a\") " pod="openshift-marketplace/redhat-marketplace-7hx5h" Dec 06 16:21:06 crc kubenswrapper[4813]: I1206 16:21:06.697364 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6724f83-f171-4e2a-a03f-4109fc38c74a-catalog-content\") pod \"redhat-marketplace-7hx5h\" (UID: \"c6724f83-f171-4e2a-a03f-4109fc38c74a\") " pod="openshift-marketplace/redhat-marketplace-7hx5h" Dec 06 16:21:06 crc kubenswrapper[4813]: I1206 16:21:06.697394 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdkx4\" (UniqueName: \"kubernetes.io/projected/c6724f83-f171-4e2a-a03f-4109fc38c74a-kube-api-access-tdkx4\") pod \"redhat-marketplace-7hx5h\" (UID: \"c6724f83-f171-4e2a-a03f-4109fc38c74a\") " pod="openshift-marketplace/redhat-marketplace-7hx5h" Dec 06 16:21:06 crc kubenswrapper[4813]: I1206 16:21:06.799332 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6724f83-f171-4e2a-a03f-4109fc38c74a-catalog-content\") pod \"redhat-marketplace-7hx5h\" (UID: \"c6724f83-f171-4e2a-a03f-4109fc38c74a\") " pod="openshift-marketplace/redhat-marketplace-7hx5h" Dec 06 16:21:06 crc kubenswrapper[4813]: I1206 16:21:06.799386 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdkx4\" (UniqueName: \"kubernetes.io/projected/c6724f83-f171-4e2a-a03f-4109fc38c74a-kube-api-access-tdkx4\") pod \"redhat-marketplace-7hx5h\" (UID: \"c6724f83-f171-4e2a-a03f-4109fc38c74a\") " pod="openshift-marketplace/redhat-marketplace-7hx5h" Dec 06 16:21:06 crc kubenswrapper[4813]: I1206 16:21:06.799551 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6724f83-f171-4e2a-a03f-4109fc38c74a-utilities\") pod \"redhat-marketplace-7hx5h\" (UID: \"c6724f83-f171-4e2a-a03f-4109fc38c74a\") " pod="openshift-marketplace/redhat-marketplace-7hx5h" Dec 06 16:21:06 crc kubenswrapper[4813]: I1206 16:21:06.800345 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6724f83-f171-4e2a-a03f-4109fc38c74a-utilities\") pod \"redhat-marketplace-7hx5h\" (UID: \"c6724f83-f171-4e2a-a03f-4109fc38c74a\") " pod="openshift-marketplace/redhat-marketplace-7hx5h" Dec 06 16:21:06 crc kubenswrapper[4813]: I1206 16:21:06.800388 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6724f83-f171-4e2a-a03f-4109fc38c74a-catalog-content\") pod \"redhat-marketplace-7hx5h\" (UID: \"c6724f83-f171-4e2a-a03f-4109fc38c74a\") " pod="openshift-marketplace/redhat-marketplace-7hx5h" Dec 06 16:21:06 crc kubenswrapper[4813]: I1206 16:21:06.832873 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdkx4\" (UniqueName: \"kubernetes.io/projected/c6724f83-f171-4e2a-a03f-4109fc38c74a-kube-api-access-tdkx4\") pod \"redhat-marketplace-7hx5h\" (UID: \"c6724f83-f171-4e2a-a03f-4109fc38c74a\") " pod="openshift-marketplace/redhat-marketplace-7hx5h" Dec 06 16:21:06 crc kubenswrapper[4813]: I1206 16:21:06.954823 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7hx5h" Dec 06 16:21:07 crc kubenswrapper[4813]: I1206 16:21:07.455239 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7hx5h"] Dec 06 16:21:08 crc kubenswrapper[4813]: I1206 16:21:08.099289 4813 generic.go:334] "Generic (PLEG): container finished" podID="c6724f83-f171-4e2a-a03f-4109fc38c74a" containerID="2ccc4599046591b389b4461eebd3b02b59ae41e3a2ad5c14952322126f81356a" exitCode=0 Dec 06 16:21:08 crc kubenswrapper[4813]: I1206 16:21:08.099520 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7hx5h" event={"ID":"c6724f83-f171-4e2a-a03f-4109fc38c74a","Type":"ContainerDied","Data":"2ccc4599046591b389b4461eebd3b02b59ae41e3a2ad5c14952322126f81356a"} Dec 06 16:21:08 crc kubenswrapper[4813]: I1206 16:21:08.099564 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7hx5h" event={"ID":"c6724f83-f171-4e2a-a03f-4109fc38c74a","Type":"ContainerStarted","Data":"740a3d56a34d88a361a188882f59103b8cae5d00a7b3034b58752adb4dd0218d"} Dec 06 16:21:09 crc kubenswrapper[4813]: I1206 16:21:09.110142 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7hx5h" event={"ID":"c6724f83-f171-4e2a-a03f-4109fc38c74a","Type":"ContainerStarted","Data":"0ac5f9a543e2bb582e6d9151bfd10b640c5f157b5c537538a55cb19b4e118c85"} Dec 06 16:21:10 crc kubenswrapper[4813]: I1206 16:21:10.121510 4813 generic.go:334] "Generic (PLEG): container finished" podID="c6724f83-f171-4e2a-a03f-4109fc38c74a" containerID="0ac5f9a543e2bb582e6d9151bfd10b640c5f157b5c537538a55cb19b4e118c85" exitCode=0 Dec 06 16:21:10 crc kubenswrapper[4813]: I1206 16:21:10.121619 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7hx5h" event={"ID":"c6724f83-f171-4e2a-a03f-4109fc38c74a","Type":"ContainerDied","Data":"0ac5f9a543e2bb582e6d9151bfd10b640c5f157b5c537538a55cb19b4e118c85"} Dec 06 16:21:11 crc kubenswrapper[4813]: I1206 16:21:11.132857 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7hx5h" event={"ID":"c6724f83-f171-4e2a-a03f-4109fc38c74a","Type":"ContainerStarted","Data":"a0a2a4a2c91452984a2213d2ba92433dc1490c67bc8b6b133e867fa90f9a4205"} Dec 06 16:21:11 crc kubenswrapper[4813]: I1206 16:21:11.154040 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7hx5h" podStartSLOduration=2.646893433 podStartE2EDuration="5.154016827s" podCreationTimestamp="2025-12-06 16:21:06 +0000 UTC" firstStartedPulling="2025-12-06 16:21:08.101571723 +0000 UTC m=+2107.992451299" lastFinishedPulling="2025-12-06 16:21:10.608695117 +0000 UTC m=+2110.499574693" observedRunningTime="2025-12-06 16:21:11.153816062 +0000 UTC m=+2111.044695648" watchObservedRunningTime="2025-12-06 16:21:11.154016827 +0000 UTC m=+2111.044896423" Dec 06 16:21:16 crc kubenswrapper[4813]: I1206 16:21:16.955725 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7hx5h" Dec 06 16:21:16 crc kubenswrapper[4813]: I1206 16:21:16.956231 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7hx5h" Dec 06 16:21:17 crc kubenswrapper[4813]: I1206 16:21:17.043193 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7hx5h" Dec 06 16:21:17 crc kubenswrapper[4813]: I1206 16:21:17.248489 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7hx5h" Dec 06 16:21:17 crc kubenswrapper[4813]: I1206 16:21:17.304606 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7hx5h"] Dec 06 16:21:19 crc kubenswrapper[4813]: I1206 16:21:19.205344 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7hx5h" podUID="c6724f83-f171-4e2a-a03f-4109fc38c74a" containerName="registry-server" containerID="cri-o://a0a2a4a2c91452984a2213d2ba92433dc1490c67bc8b6b133e867fa90f9a4205" gracePeriod=2 Dec 06 16:21:19 crc kubenswrapper[4813]: I1206 16:21:19.675526 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7hx5h" Dec 06 16:21:19 crc kubenswrapper[4813]: I1206 16:21:19.783369 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6724f83-f171-4e2a-a03f-4109fc38c74a-catalog-content\") pod \"c6724f83-f171-4e2a-a03f-4109fc38c74a\" (UID: \"c6724f83-f171-4e2a-a03f-4109fc38c74a\") " Dec 06 16:21:19 crc kubenswrapper[4813]: I1206 16:21:19.783554 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6724f83-f171-4e2a-a03f-4109fc38c74a-utilities\") pod \"c6724f83-f171-4e2a-a03f-4109fc38c74a\" (UID: \"c6724f83-f171-4e2a-a03f-4109fc38c74a\") " Dec 06 16:21:19 crc kubenswrapper[4813]: I1206 16:21:19.783683 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdkx4\" (UniqueName: \"kubernetes.io/projected/c6724f83-f171-4e2a-a03f-4109fc38c74a-kube-api-access-tdkx4\") pod \"c6724f83-f171-4e2a-a03f-4109fc38c74a\" (UID: \"c6724f83-f171-4e2a-a03f-4109fc38c74a\") " Dec 06 16:21:19 crc kubenswrapper[4813]: I1206 16:21:19.784077 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6724f83-f171-4e2a-a03f-4109fc38c74a-utilities" (OuterVolumeSpecName: "utilities") pod "c6724f83-f171-4e2a-a03f-4109fc38c74a" (UID: "c6724f83-f171-4e2a-a03f-4109fc38c74a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:21:19 crc kubenswrapper[4813]: I1206 16:21:19.794531 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6724f83-f171-4e2a-a03f-4109fc38c74a-kube-api-access-tdkx4" (OuterVolumeSpecName: "kube-api-access-tdkx4") pod "c6724f83-f171-4e2a-a03f-4109fc38c74a" (UID: "c6724f83-f171-4e2a-a03f-4109fc38c74a"). InnerVolumeSpecName "kube-api-access-tdkx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:21:19 crc kubenswrapper[4813]: I1206 16:21:19.805473 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6724f83-f171-4e2a-a03f-4109fc38c74a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c6724f83-f171-4e2a-a03f-4109fc38c74a" (UID: "c6724f83-f171-4e2a-a03f-4109fc38c74a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:21:19 crc kubenswrapper[4813]: I1206 16:21:19.892983 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6724f83-f171-4e2a-a03f-4109fc38c74a-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 16:21:19 crc kubenswrapper[4813]: I1206 16:21:19.893022 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdkx4\" (UniqueName: \"kubernetes.io/projected/c6724f83-f171-4e2a-a03f-4109fc38c74a-kube-api-access-tdkx4\") on node \"crc\" DevicePath \"\"" Dec 06 16:21:19 crc kubenswrapper[4813]: I1206 16:21:19.893035 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6724f83-f171-4e2a-a03f-4109fc38c74a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 16:21:20 crc kubenswrapper[4813]: I1206 16:21:20.216537 4813 generic.go:334] "Generic (PLEG): container finished" podID="c6724f83-f171-4e2a-a03f-4109fc38c74a" containerID="a0a2a4a2c91452984a2213d2ba92433dc1490c67bc8b6b133e867fa90f9a4205" exitCode=0 Dec 06 16:21:20 crc kubenswrapper[4813]: I1206 16:21:20.216578 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7hx5h" event={"ID":"c6724f83-f171-4e2a-a03f-4109fc38c74a","Type":"ContainerDied","Data":"a0a2a4a2c91452984a2213d2ba92433dc1490c67bc8b6b133e867fa90f9a4205"} Dec 06 16:21:20 crc kubenswrapper[4813]: I1206 16:21:20.216604 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7hx5h" event={"ID":"c6724f83-f171-4e2a-a03f-4109fc38c74a","Type":"ContainerDied","Data":"740a3d56a34d88a361a188882f59103b8cae5d00a7b3034b58752adb4dd0218d"} Dec 06 16:21:20 crc kubenswrapper[4813]: I1206 16:21:20.216619 4813 scope.go:117] "RemoveContainer" containerID="a0a2a4a2c91452984a2213d2ba92433dc1490c67bc8b6b133e867fa90f9a4205" Dec 06 16:21:20 crc kubenswrapper[4813]: I1206 16:21:20.216625 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7hx5h" Dec 06 16:21:20 crc kubenswrapper[4813]: I1206 16:21:20.239020 4813 scope.go:117] "RemoveContainer" containerID="0ac5f9a543e2bb582e6d9151bfd10b640c5f157b5c537538a55cb19b4e118c85" Dec 06 16:21:20 crc kubenswrapper[4813]: I1206 16:21:20.275533 4813 scope.go:117] "RemoveContainer" containerID="2ccc4599046591b389b4461eebd3b02b59ae41e3a2ad5c14952322126f81356a" Dec 06 16:21:20 crc kubenswrapper[4813]: I1206 16:21:20.275672 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7hx5h"] Dec 06 16:21:20 crc kubenswrapper[4813]: I1206 16:21:20.284045 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7hx5h"] Dec 06 16:21:20 crc kubenswrapper[4813]: I1206 16:21:20.309928 4813 scope.go:117] "RemoveContainer" containerID="a0a2a4a2c91452984a2213d2ba92433dc1490c67bc8b6b133e867fa90f9a4205" Dec 06 16:21:20 crc kubenswrapper[4813]: E1206 16:21:20.311046 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0a2a4a2c91452984a2213d2ba92433dc1490c67bc8b6b133e867fa90f9a4205\": container with ID starting with a0a2a4a2c91452984a2213d2ba92433dc1490c67bc8b6b133e867fa90f9a4205 not found: ID does not exist" containerID="a0a2a4a2c91452984a2213d2ba92433dc1490c67bc8b6b133e867fa90f9a4205" Dec 06 16:21:20 crc kubenswrapper[4813]: I1206 16:21:20.311082 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0a2a4a2c91452984a2213d2ba92433dc1490c67bc8b6b133e867fa90f9a4205"} err="failed to get container status \"a0a2a4a2c91452984a2213d2ba92433dc1490c67bc8b6b133e867fa90f9a4205\": rpc error: code = NotFound desc = could not find container \"a0a2a4a2c91452984a2213d2ba92433dc1490c67bc8b6b133e867fa90f9a4205\": container with ID starting with a0a2a4a2c91452984a2213d2ba92433dc1490c67bc8b6b133e867fa90f9a4205 not found: ID does not exist" Dec 06 16:21:20 crc kubenswrapper[4813]: I1206 16:21:20.311106 4813 scope.go:117] "RemoveContainer" containerID="0ac5f9a543e2bb582e6d9151bfd10b640c5f157b5c537538a55cb19b4e118c85" Dec 06 16:21:20 crc kubenswrapper[4813]: E1206 16:21:20.315455 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ac5f9a543e2bb582e6d9151bfd10b640c5f157b5c537538a55cb19b4e118c85\": container with ID starting with 0ac5f9a543e2bb582e6d9151bfd10b640c5f157b5c537538a55cb19b4e118c85 not found: ID does not exist" containerID="0ac5f9a543e2bb582e6d9151bfd10b640c5f157b5c537538a55cb19b4e118c85" Dec 06 16:21:20 crc kubenswrapper[4813]: I1206 16:21:20.315499 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ac5f9a543e2bb582e6d9151bfd10b640c5f157b5c537538a55cb19b4e118c85"} err="failed to get container status \"0ac5f9a543e2bb582e6d9151bfd10b640c5f157b5c537538a55cb19b4e118c85\": rpc error: code = NotFound desc = could not find container \"0ac5f9a543e2bb582e6d9151bfd10b640c5f157b5c537538a55cb19b4e118c85\": container with ID starting with 0ac5f9a543e2bb582e6d9151bfd10b640c5f157b5c537538a55cb19b4e118c85 not found: ID does not exist" Dec 06 16:21:20 crc kubenswrapper[4813]: I1206 16:21:20.315538 4813 scope.go:117] "RemoveContainer" containerID="2ccc4599046591b389b4461eebd3b02b59ae41e3a2ad5c14952322126f81356a" Dec 06 16:21:20 crc kubenswrapper[4813]: E1206 16:21:20.315857 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ccc4599046591b389b4461eebd3b02b59ae41e3a2ad5c14952322126f81356a\": container with ID starting with 2ccc4599046591b389b4461eebd3b02b59ae41e3a2ad5c14952322126f81356a not found: ID does not exist" containerID="2ccc4599046591b389b4461eebd3b02b59ae41e3a2ad5c14952322126f81356a" Dec 06 16:21:20 crc kubenswrapper[4813]: I1206 16:21:20.315883 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ccc4599046591b389b4461eebd3b02b59ae41e3a2ad5c14952322126f81356a"} err="failed to get container status \"2ccc4599046591b389b4461eebd3b02b59ae41e3a2ad5c14952322126f81356a\": rpc error: code = NotFound desc = could not find container \"2ccc4599046591b389b4461eebd3b02b59ae41e3a2ad5c14952322126f81356a\": container with ID starting with 2ccc4599046591b389b4461eebd3b02b59ae41e3a2ad5c14952322126f81356a not found: ID does not exist" Dec 06 16:21:20 crc kubenswrapper[4813]: I1206 16:21:20.498184 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6724f83-f171-4e2a-a03f-4109fc38c74a" path="/var/lib/kubelet/pods/c6724f83-f171-4e2a-a03f-4109fc38c74a/volumes" Dec 06 16:21:40 crc kubenswrapper[4813]: I1206 16:21:40.361435 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cqw6q"] Dec 06 16:21:40 crc kubenswrapper[4813]: E1206 16:21:40.362402 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6724f83-f171-4e2a-a03f-4109fc38c74a" containerName="registry-server" Dec 06 16:21:40 crc kubenswrapper[4813]: I1206 16:21:40.362418 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6724f83-f171-4e2a-a03f-4109fc38c74a" containerName="registry-server" Dec 06 16:21:40 crc kubenswrapper[4813]: E1206 16:21:40.362435 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6724f83-f171-4e2a-a03f-4109fc38c74a" containerName="extract-utilities" Dec 06 16:21:40 crc kubenswrapper[4813]: I1206 16:21:40.362444 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6724f83-f171-4e2a-a03f-4109fc38c74a" containerName="extract-utilities" Dec 06 16:21:40 crc kubenswrapper[4813]: E1206 16:21:40.362489 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6724f83-f171-4e2a-a03f-4109fc38c74a" containerName="extract-content" Dec 06 16:21:40 crc kubenswrapper[4813]: I1206 16:21:40.362499 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6724f83-f171-4e2a-a03f-4109fc38c74a" containerName="extract-content" Dec 06 16:21:40 crc kubenswrapper[4813]: I1206 16:21:40.362760 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6724f83-f171-4e2a-a03f-4109fc38c74a" containerName="registry-server" Dec 06 16:21:40 crc kubenswrapper[4813]: I1206 16:21:40.364507 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cqw6q" Dec 06 16:21:40 crc kubenswrapper[4813]: I1206 16:21:40.385800 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cqw6q"] Dec 06 16:21:40 crc kubenswrapper[4813]: I1206 16:21:40.481799 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8d5a581-2850-412a-95f1-afd1a7db1407-catalog-content\") pod \"community-operators-cqw6q\" (UID: \"c8d5a581-2850-412a-95f1-afd1a7db1407\") " pod="openshift-marketplace/community-operators-cqw6q" Dec 06 16:21:40 crc kubenswrapper[4813]: I1206 16:21:40.482116 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfmbs\" (UniqueName: \"kubernetes.io/projected/c8d5a581-2850-412a-95f1-afd1a7db1407-kube-api-access-lfmbs\") pod \"community-operators-cqw6q\" (UID: \"c8d5a581-2850-412a-95f1-afd1a7db1407\") " pod="openshift-marketplace/community-operators-cqw6q" Dec 06 16:21:40 crc kubenswrapper[4813]: I1206 16:21:40.482137 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8d5a581-2850-412a-95f1-afd1a7db1407-utilities\") pod \"community-operators-cqw6q\" (UID: \"c8d5a581-2850-412a-95f1-afd1a7db1407\") " pod="openshift-marketplace/community-operators-cqw6q" Dec 06 16:21:40 crc kubenswrapper[4813]: I1206 16:21:40.583284 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8d5a581-2850-412a-95f1-afd1a7db1407-catalog-content\") pod \"community-operators-cqw6q\" (UID: \"c8d5a581-2850-412a-95f1-afd1a7db1407\") " pod="openshift-marketplace/community-operators-cqw6q" Dec 06 16:21:40 crc kubenswrapper[4813]: I1206 16:21:40.583460 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfmbs\" (UniqueName: \"kubernetes.io/projected/c8d5a581-2850-412a-95f1-afd1a7db1407-kube-api-access-lfmbs\") pod \"community-operators-cqw6q\" (UID: \"c8d5a581-2850-412a-95f1-afd1a7db1407\") " pod="openshift-marketplace/community-operators-cqw6q" Dec 06 16:21:40 crc kubenswrapper[4813]: I1206 16:21:40.583493 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8d5a581-2850-412a-95f1-afd1a7db1407-utilities\") pod \"community-operators-cqw6q\" (UID: \"c8d5a581-2850-412a-95f1-afd1a7db1407\") " pod="openshift-marketplace/community-operators-cqw6q" Dec 06 16:21:40 crc kubenswrapper[4813]: I1206 16:21:40.584523 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8d5a581-2850-412a-95f1-afd1a7db1407-utilities\") pod \"community-operators-cqw6q\" (UID: \"c8d5a581-2850-412a-95f1-afd1a7db1407\") " pod="openshift-marketplace/community-operators-cqw6q" Dec 06 16:21:40 crc kubenswrapper[4813]: I1206 16:21:40.585192 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8d5a581-2850-412a-95f1-afd1a7db1407-catalog-content\") pod \"community-operators-cqw6q\" (UID: \"c8d5a581-2850-412a-95f1-afd1a7db1407\") " pod="openshift-marketplace/community-operators-cqw6q" Dec 06 16:21:40 crc kubenswrapper[4813]: I1206 16:21:40.609528 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfmbs\" (UniqueName: \"kubernetes.io/projected/c8d5a581-2850-412a-95f1-afd1a7db1407-kube-api-access-lfmbs\") pod \"community-operators-cqw6q\" (UID: \"c8d5a581-2850-412a-95f1-afd1a7db1407\") " pod="openshift-marketplace/community-operators-cqw6q" Dec 06 16:21:40 crc kubenswrapper[4813]: I1206 16:21:40.715438 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cqw6q" Dec 06 16:21:41 crc kubenswrapper[4813]: I1206 16:21:41.485236 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cqw6q"] Dec 06 16:21:42 crc kubenswrapper[4813]: I1206 16:21:42.441900 4813 generic.go:334] "Generic (PLEG): container finished" podID="c8d5a581-2850-412a-95f1-afd1a7db1407" containerID="d59fff2e90e3d4c7132f5bfffa206fef124aa948012334234225d49ead2bac26" exitCode=0 Dec 06 16:21:42 crc kubenswrapper[4813]: I1206 16:21:42.441965 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cqw6q" event={"ID":"c8d5a581-2850-412a-95f1-afd1a7db1407","Type":"ContainerDied","Data":"d59fff2e90e3d4c7132f5bfffa206fef124aa948012334234225d49ead2bac26"} Dec 06 16:21:42 crc kubenswrapper[4813]: I1206 16:21:42.442513 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cqw6q" event={"ID":"c8d5a581-2850-412a-95f1-afd1a7db1407","Type":"ContainerStarted","Data":"2f316ac41d314e009127cc0beef3e383752c87b104c67061cfa558f69fce8dec"} Dec 06 16:21:43 crc kubenswrapper[4813]: I1206 16:21:43.455150 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cqw6q" event={"ID":"c8d5a581-2850-412a-95f1-afd1a7db1407","Type":"ContainerStarted","Data":"2192c6ea68e1cc971ce43c9e6dbe79dbe5b003ae9e66d2b020ab39003dc557a2"} Dec 06 16:21:44 crc kubenswrapper[4813]: I1206 16:21:44.468440 4813 generic.go:334] "Generic (PLEG): container finished" podID="c8d5a581-2850-412a-95f1-afd1a7db1407" containerID="2192c6ea68e1cc971ce43c9e6dbe79dbe5b003ae9e66d2b020ab39003dc557a2" exitCode=0 Dec 06 16:21:44 crc kubenswrapper[4813]: I1206 16:21:44.468604 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cqw6q" event={"ID":"c8d5a581-2850-412a-95f1-afd1a7db1407","Type":"ContainerDied","Data":"2192c6ea68e1cc971ce43c9e6dbe79dbe5b003ae9e66d2b020ab39003dc557a2"} Dec 06 16:21:45 crc kubenswrapper[4813]: I1206 16:21:45.483436 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cqw6q" event={"ID":"c8d5a581-2850-412a-95f1-afd1a7db1407","Type":"ContainerStarted","Data":"2e161f58f355a77356ba7de8d3e86c70a3a80249ce774df594fa5def572287d3"} Dec 06 16:21:45 crc kubenswrapper[4813]: I1206 16:21:45.503360 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cqw6q" podStartSLOduration=2.969249644 podStartE2EDuration="5.503343393s" podCreationTimestamp="2025-12-06 16:21:40 +0000 UTC" firstStartedPulling="2025-12-06 16:21:42.444011438 +0000 UTC m=+2142.334891004" lastFinishedPulling="2025-12-06 16:21:44.978105157 +0000 UTC m=+2144.868984753" observedRunningTime="2025-12-06 16:21:45.500823378 +0000 UTC m=+2145.391702964" watchObservedRunningTime="2025-12-06 16:21:45.503343393 +0000 UTC m=+2145.394222979" Dec 06 16:21:50 crc kubenswrapper[4813]: I1206 16:21:50.716035 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cqw6q" Dec 06 16:21:50 crc kubenswrapper[4813]: I1206 16:21:50.716642 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cqw6q" Dec 06 16:21:50 crc kubenswrapper[4813]: I1206 16:21:50.772814 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cqw6q" Dec 06 16:21:51 crc kubenswrapper[4813]: I1206 16:21:51.608253 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cqw6q" Dec 06 16:21:51 crc kubenswrapper[4813]: I1206 16:21:51.692065 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cqw6q"] Dec 06 16:21:53 crc kubenswrapper[4813]: I1206 16:21:53.577454 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cqw6q" podUID="c8d5a581-2850-412a-95f1-afd1a7db1407" containerName="registry-server" containerID="cri-o://2e161f58f355a77356ba7de8d3e86c70a3a80249ce774df594fa5def572287d3" gracePeriod=2 Dec 06 16:21:54 crc kubenswrapper[4813]: I1206 16:21:54.129021 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cqw6q" Dec 06 16:21:54 crc kubenswrapper[4813]: I1206 16:21:54.308569 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8d5a581-2850-412a-95f1-afd1a7db1407-catalog-content\") pod \"c8d5a581-2850-412a-95f1-afd1a7db1407\" (UID: \"c8d5a581-2850-412a-95f1-afd1a7db1407\") " Dec 06 16:21:54 crc kubenswrapper[4813]: I1206 16:21:54.309067 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfmbs\" (UniqueName: \"kubernetes.io/projected/c8d5a581-2850-412a-95f1-afd1a7db1407-kube-api-access-lfmbs\") pod \"c8d5a581-2850-412a-95f1-afd1a7db1407\" (UID: \"c8d5a581-2850-412a-95f1-afd1a7db1407\") " Dec 06 16:21:54 crc kubenswrapper[4813]: I1206 16:21:54.309222 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8d5a581-2850-412a-95f1-afd1a7db1407-utilities\") pod \"c8d5a581-2850-412a-95f1-afd1a7db1407\" (UID: \"c8d5a581-2850-412a-95f1-afd1a7db1407\") " Dec 06 16:21:54 crc kubenswrapper[4813]: I1206 16:21:54.311514 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8d5a581-2850-412a-95f1-afd1a7db1407-utilities" (OuterVolumeSpecName: "utilities") pod "c8d5a581-2850-412a-95f1-afd1a7db1407" (UID: "c8d5a581-2850-412a-95f1-afd1a7db1407"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:21:54 crc kubenswrapper[4813]: I1206 16:21:54.319520 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8d5a581-2850-412a-95f1-afd1a7db1407-kube-api-access-lfmbs" (OuterVolumeSpecName: "kube-api-access-lfmbs") pod "c8d5a581-2850-412a-95f1-afd1a7db1407" (UID: "c8d5a581-2850-412a-95f1-afd1a7db1407"). InnerVolumeSpecName "kube-api-access-lfmbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:21:54 crc kubenswrapper[4813]: I1206 16:21:54.368042 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8d5a581-2850-412a-95f1-afd1a7db1407-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c8d5a581-2850-412a-95f1-afd1a7db1407" (UID: "c8d5a581-2850-412a-95f1-afd1a7db1407"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:21:54 crc kubenswrapper[4813]: I1206 16:21:54.412074 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8d5a581-2850-412a-95f1-afd1a7db1407-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 16:21:54 crc kubenswrapper[4813]: I1206 16:21:54.412320 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8d5a581-2850-412a-95f1-afd1a7db1407-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 16:21:54 crc kubenswrapper[4813]: I1206 16:21:54.412332 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfmbs\" (UniqueName: \"kubernetes.io/projected/c8d5a581-2850-412a-95f1-afd1a7db1407-kube-api-access-lfmbs\") on node \"crc\" DevicePath \"\"" Dec 06 16:21:54 crc kubenswrapper[4813]: I1206 16:21:54.651867 4813 generic.go:334] "Generic (PLEG): container finished" podID="c8d5a581-2850-412a-95f1-afd1a7db1407" containerID="2e161f58f355a77356ba7de8d3e86c70a3a80249ce774df594fa5def572287d3" exitCode=0 Dec 06 16:21:54 crc kubenswrapper[4813]: I1206 16:21:54.651910 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cqw6q" event={"ID":"c8d5a581-2850-412a-95f1-afd1a7db1407","Type":"ContainerDied","Data":"2e161f58f355a77356ba7de8d3e86c70a3a80249ce774df594fa5def572287d3"} Dec 06 16:21:54 crc kubenswrapper[4813]: I1206 16:21:54.651940 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cqw6q" event={"ID":"c8d5a581-2850-412a-95f1-afd1a7db1407","Type":"ContainerDied","Data":"2f316ac41d314e009127cc0beef3e383752c87b104c67061cfa558f69fce8dec"} Dec 06 16:21:54 crc kubenswrapper[4813]: I1206 16:21:54.651942 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cqw6q" Dec 06 16:21:54 crc kubenswrapper[4813]: I1206 16:21:54.651958 4813 scope.go:117] "RemoveContainer" containerID="2e161f58f355a77356ba7de8d3e86c70a3a80249ce774df594fa5def572287d3" Dec 06 16:21:54 crc kubenswrapper[4813]: I1206 16:21:54.679980 4813 scope.go:117] "RemoveContainer" containerID="2192c6ea68e1cc971ce43c9e6dbe79dbe5b003ae9e66d2b020ab39003dc557a2" Dec 06 16:21:54 crc kubenswrapper[4813]: I1206 16:21:54.696602 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cqw6q"] Dec 06 16:21:54 crc kubenswrapper[4813]: I1206 16:21:54.705621 4813 scope.go:117] "RemoveContainer" containerID="d59fff2e90e3d4c7132f5bfffa206fef124aa948012334234225d49ead2bac26" Dec 06 16:21:54 crc kubenswrapper[4813]: I1206 16:21:54.711332 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cqw6q"] Dec 06 16:21:54 crc kubenswrapper[4813]: I1206 16:21:54.755178 4813 scope.go:117] "RemoveContainer" containerID="2e161f58f355a77356ba7de8d3e86c70a3a80249ce774df594fa5def572287d3" Dec 06 16:21:54 crc kubenswrapper[4813]: E1206 16:21:54.755778 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e161f58f355a77356ba7de8d3e86c70a3a80249ce774df594fa5def572287d3\": container with ID starting with 2e161f58f355a77356ba7de8d3e86c70a3a80249ce774df594fa5def572287d3 not found: ID does not exist" containerID="2e161f58f355a77356ba7de8d3e86c70a3a80249ce774df594fa5def572287d3" Dec 06 16:21:54 crc kubenswrapper[4813]: I1206 16:21:54.755819 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e161f58f355a77356ba7de8d3e86c70a3a80249ce774df594fa5def572287d3"} err="failed to get container status \"2e161f58f355a77356ba7de8d3e86c70a3a80249ce774df594fa5def572287d3\": rpc error: code = NotFound desc = could not find container \"2e161f58f355a77356ba7de8d3e86c70a3a80249ce774df594fa5def572287d3\": container with ID starting with 2e161f58f355a77356ba7de8d3e86c70a3a80249ce774df594fa5def572287d3 not found: ID does not exist" Dec 06 16:21:54 crc kubenswrapper[4813]: I1206 16:21:54.755845 4813 scope.go:117] "RemoveContainer" containerID="2192c6ea68e1cc971ce43c9e6dbe79dbe5b003ae9e66d2b020ab39003dc557a2" Dec 06 16:21:54 crc kubenswrapper[4813]: E1206 16:21:54.756241 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2192c6ea68e1cc971ce43c9e6dbe79dbe5b003ae9e66d2b020ab39003dc557a2\": container with ID starting with 2192c6ea68e1cc971ce43c9e6dbe79dbe5b003ae9e66d2b020ab39003dc557a2 not found: ID does not exist" containerID="2192c6ea68e1cc971ce43c9e6dbe79dbe5b003ae9e66d2b020ab39003dc557a2" Dec 06 16:21:54 crc kubenswrapper[4813]: I1206 16:21:54.756370 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2192c6ea68e1cc971ce43c9e6dbe79dbe5b003ae9e66d2b020ab39003dc557a2"} err="failed to get container status \"2192c6ea68e1cc971ce43c9e6dbe79dbe5b003ae9e66d2b020ab39003dc557a2\": rpc error: code = NotFound desc = could not find container \"2192c6ea68e1cc971ce43c9e6dbe79dbe5b003ae9e66d2b020ab39003dc557a2\": container with ID starting with 2192c6ea68e1cc971ce43c9e6dbe79dbe5b003ae9e66d2b020ab39003dc557a2 not found: ID does not exist" Dec 06 16:21:54 crc kubenswrapper[4813]: I1206 16:21:54.756388 4813 scope.go:117] "RemoveContainer" containerID="d59fff2e90e3d4c7132f5bfffa206fef124aa948012334234225d49ead2bac26" Dec 06 16:21:54 crc kubenswrapper[4813]: E1206 16:21:54.756791 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d59fff2e90e3d4c7132f5bfffa206fef124aa948012334234225d49ead2bac26\": container with ID starting with d59fff2e90e3d4c7132f5bfffa206fef124aa948012334234225d49ead2bac26 not found: ID does not exist" containerID="d59fff2e90e3d4c7132f5bfffa206fef124aa948012334234225d49ead2bac26" Dec 06 16:21:54 crc kubenswrapper[4813]: I1206 16:21:54.756818 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d59fff2e90e3d4c7132f5bfffa206fef124aa948012334234225d49ead2bac26"} err="failed to get container status \"d59fff2e90e3d4c7132f5bfffa206fef124aa948012334234225d49ead2bac26\": rpc error: code = NotFound desc = could not find container \"d59fff2e90e3d4c7132f5bfffa206fef124aa948012334234225d49ead2bac26\": container with ID starting with d59fff2e90e3d4c7132f5bfffa206fef124aa948012334234225d49ead2bac26 not found: ID does not exist" Dec 06 16:21:56 crc kubenswrapper[4813]: I1206 16:21:56.497792 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8d5a581-2850-412a-95f1-afd1a7db1407" path="/var/lib/kubelet/pods/c8d5a581-2850-412a-95f1-afd1a7db1407/volumes" Dec 06 16:22:49 crc kubenswrapper[4813]: I1206 16:22:49.427808 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:22:49 crc kubenswrapper[4813]: I1206 16:22:49.428482 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:23:04 crc kubenswrapper[4813]: I1206 16:23:04.133133 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bqsvv"] Dec 06 16:23:04 crc kubenswrapper[4813]: E1206 16:23:04.133960 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8d5a581-2850-412a-95f1-afd1a7db1407" containerName="extract-content" Dec 06 16:23:04 crc kubenswrapper[4813]: I1206 16:23:04.133973 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8d5a581-2850-412a-95f1-afd1a7db1407" containerName="extract-content" Dec 06 16:23:04 crc kubenswrapper[4813]: E1206 16:23:04.134000 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8d5a581-2850-412a-95f1-afd1a7db1407" containerName="registry-server" Dec 06 16:23:04 crc kubenswrapper[4813]: I1206 16:23:04.134007 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8d5a581-2850-412a-95f1-afd1a7db1407" containerName="registry-server" Dec 06 16:23:04 crc kubenswrapper[4813]: E1206 16:23:04.134024 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8d5a581-2850-412a-95f1-afd1a7db1407" containerName="extract-utilities" Dec 06 16:23:04 crc kubenswrapper[4813]: I1206 16:23:04.134032 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8d5a581-2850-412a-95f1-afd1a7db1407" containerName="extract-utilities" Dec 06 16:23:04 crc kubenswrapper[4813]: I1206 16:23:04.134251 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8d5a581-2850-412a-95f1-afd1a7db1407" containerName="registry-server" Dec 06 16:23:04 crc kubenswrapper[4813]: I1206 16:23:04.135663 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bqsvv" Dec 06 16:23:04 crc kubenswrapper[4813]: I1206 16:23:04.151459 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtzqm\" (UniqueName: \"kubernetes.io/projected/7458cfeb-9419-45cb-adf9-05c35f988b3e-kube-api-access-vtzqm\") pod \"redhat-operators-bqsvv\" (UID: \"7458cfeb-9419-45cb-adf9-05c35f988b3e\") " pod="openshift-marketplace/redhat-operators-bqsvv" Dec 06 16:23:04 crc kubenswrapper[4813]: I1206 16:23:04.151711 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7458cfeb-9419-45cb-adf9-05c35f988b3e-utilities\") pod \"redhat-operators-bqsvv\" (UID: \"7458cfeb-9419-45cb-adf9-05c35f988b3e\") " pod="openshift-marketplace/redhat-operators-bqsvv" Dec 06 16:23:04 crc kubenswrapper[4813]: I1206 16:23:04.151845 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7458cfeb-9419-45cb-adf9-05c35f988b3e-catalog-content\") pod \"redhat-operators-bqsvv\" (UID: \"7458cfeb-9419-45cb-adf9-05c35f988b3e\") " pod="openshift-marketplace/redhat-operators-bqsvv" Dec 06 16:23:04 crc kubenswrapper[4813]: I1206 16:23:04.157663 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bqsvv"] Dec 06 16:23:04 crc kubenswrapper[4813]: I1206 16:23:04.252818 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtzqm\" (UniqueName: \"kubernetes.io/projected/7458cfeb-9419-45cb-adf9-05c35f988b3e-kube-api-access-vtzqm\") pod \"redhat-operators-bqsvv\" (UID: \"7458cfeb-9419-45cb-adf9-05c35f988b3e\") " pod="openshift-marketplace/redhat-operators-bqsvv" Dec 06 16:23:04 crc kubenswrapper[4813]: I1206 16:23:04.252875 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7458cfeb-9419-45cb-adf9-05c35f988b3e-utilities\") pod \"redhat-operators-bqsvv\" (UID: \"7458cfeb-9419-45cb-adf9-05c35f988b3e\") " pod="openshift-marketplace/redhat-operators-bqsvv" Dec 06 16:23:04 crc kubenswrapper[4813]: I1206 16:23:04.252979 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7458cfeb-9419-45cb-adf9-05c35f988b3e-catalog-content\") pod \"redhat-operators-bqsvv\" (UID: \"7458cfeb-9419-45cb-adf9-05c35f988b3e\") " pod="openshift-marketplace/redhat-operators-bqsvv" Dec 06 16:23:04 crc kubenswrapper[4813]: I1206 16:23:04.253675 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7458cfeb-9419-45cb-adf9-05c35f988b3e-catalog-content\") pod \"redhat-operators-bqsvv\" (UID: \"7458cfeb-9419-45cb-adf9-05c35f988b3e\") " pod="openshift-marketplace/redhat-operators-bqsvv" Dec 06 16:23:04 crc kubenswrapper[4813]: I1206 16:23:04.253906 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7458cfeb-9419-45cb-adf9-05c35f988b3e-utilities\") pod \"redhat-operators-bqsvv\" (UID: \"7458cfeb-9419-45cb-adf9-05c35f988b3e\") " pod="openshift-marketplace/redhat-operators-bqsvv" Dec 06 16:23:04 crc kubenswrapper[4813]: I1206 16:23:04.278289 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtzqm\" (UniqueName: \"kubernetes.io/projected/7458cfeb-9419-45cb-adf9-05c35f988b3e-kube-api-access-vtzqm\") pod \"redhat-operators-bqsvv\" (UID: \"7458cfeb-9419-45cb-adf9-05c35f988b3e\") " pod="openshift-marketplace/redhat-operators-bqsvv" Dec 06 16:23:04 crc kubenswrapper[4813]: I1206 16:23:04.455078 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bqsvv" Dec 06 16:23:04 crc kubenswrapper[4813]: I1206 16:23:04.942589 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bqsvv"] Dec 06 16:23:05 crc kubenswrapper[4813]: I1206 16:23:05.482116 4813 generic.go:334] "Generic (PLEG): container finished" podID="7458cfeb-9419-45cb-adf9-05c35f988b3e" containerID="72f4f99fa784c6935864342b4e9a1f216b6626762f4dd51057a8ea202ad7f645" exitCode=0 Dec 06 16:23:05 crc kubenswrapper[4813]: I1206 16:23:05.482428 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bqsvv" event={"ID":"7458cfeb-9419-45cb-adf9-05c35f988b3e","Type":"ContainerDied","Data":"72f4f99fa784c6935864342b4e9a1f216b6626762f4dd51057a8ea202ad7f645"} Dec 06 16:23:05 crc kubenswrapper[4813]: I1206 16:23:05.482517 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bqsvv" event={"ID":"7458cfeb-9419-45cb-adf9-05c35f988b3e","Type":"ContainerStarted","Data":"71db26c103397ddcd25c33fceacaa0d91436c852aff7794316fbb4539e1f5791"} Dec 06 16:23:05 crc kubenswrapper[4813]: I1206 16:23:05.483742 4813 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 16:23:06 crc kubenswrapper[4813]: I1206 16:23:06.516448 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bqsvv" event={"ID":"7458cfeb-9419-45cb-adf9-05c35f988b3e","Type":"ContainerStarted","Data":"39faa1f9faeadcb062a911288bac9f85c5f5515527c46edb59329a12c8224f17"} Dec 06 16:23:09 crc kubenswrapper[4813]: I1206 16:23:09.574940 4813 generic.go:334] "Generic (PLEG): container finished" podID="7458cfeb-9419-45cb-adf9-05c35f988b3e" containerID="39faa1f9faeadcb062a911288bac9f85c5f5515527c46edb59329a12c8224f17" exitCode=0 Dec 06 16:23:09 crc kubenswrapper[4813]: I1206 16:23:09.575071 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bqsvv" event={"ID":"7458cfeb-9419-45cb-adf9-05c35f988b3e","Type":"ContainerDied","Data":"39faa1f9faeadcb062a911288bac9f85c5f5515527c46edb59329a12c8224f17"} Dec 06 16:23:10 crc kubenswrapper[4813]: I1206 16:23:10.590094 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bqsvv" event={"ID":"7458cfeb-9419-45cb-adf9-05c35f988b3e","Type":"ContainerStarted","Data":"a7deb66f46515059fc25463214cea26f2d7f26e7b32559ac73387eea17ae7159"} Dec 06 16:23:10 crc kubenswrapper[4813]: I1206 16:23:10.619705 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bqsvv" podStartSLOduration=2.054264091 podStartE2EDuration="6.619684218s" podCreationTimestamp="2025-12-06 16:23:04 +0000 UTC" firstStartedPulling="2025-12-06 16:23:05.483513195 +0000 UTC m=+2225.374392771" lastFinishedPulling="2025-12-06 16:23:10.048933282 +0000 UTC m=+2229.939812898" observedRunningTime="2025-12-06 16:23:10.612836318 +0000 UTC m=+2230.503715914" watchObservedRunningTime="2025-12-06 16:23:10.619684218 +0000 UTC m=+2230.510563814" Dec 06 16:23:14 crc kubenswrapper[4813]: I1206 16:23:14.455177 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bqsvv" Dec 06 16:23:14 crc kubenswrapper[4813]: I1206 16:23:14.455918 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bqsvv" Dec 06 16:23:15 crc kubenswrapper[4813]: I1206 16:23:15.523452 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bqsvv" podUID="7458cfeb-9419-45cb-adf9-05c35f988b3e" containerName="registry-server" probeResult="failure" output=< Dec 06 16:23:15 crc kubenswrapper[4813]: timeout: failed to connect service ":50051" within 1s Dec 06 16:23:15 crc kubenswrapper[4813]: > Dec 06 16:23:19 crc kubenswrapper[4813]: I1206 16:23:19.428736 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:23:19 crc kubenswrapper[4813]: I1206 16:23:19.429648 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:23:24 crc kubenswrapper[4813]: I1206 16:23:24.524238 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bqsvv" Dec 06 16:23:24 crc kubenswrapper[4813]: I1206 16:23:24.619226 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bqsvv" Dec 06 16:23:24 crc kubenswrapper[4813]: I1206 16:23:24.779925 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bqsvv"] Dec 06 16:23:25 crc kubenswrapper[4813]: I1206 16:23:25.734570 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bqsvv" podUID="7458cfeb-9419-45cb-adf9-05c35f988b3e" containerName="registry-server" containerID="cri-o://a7deb66f46515059fc25463214cea26f2d7f26e7b32559ac73387eea17ae7159" gracePeriod=2 Dec 06 16:23:26 crc kubenswrapper[4813]: I1206 16:23:26.272535 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bqsvv" Dec 06 16:23:26 crc kubenswrapper[4813]: I1206 16:23:26.354361 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7458cfeb-9419-45cb-adf9-05c35f988b3e-catalog-content\") pod \"7458cfeb-9419-45cb-adf9-05c35f988b3e\" (UID: \"7458cfeb-9419-45cb-adf9-05c35f988b3e\") " Dec 06 16:23:26 crc kubenswrapper[4813]: I1206 16:23:26.354568 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtzqm\" (UniqueName: \"kubernetes.io/projected/7458cfeb-9419-45cb-adf9-05c35f988b3e-kube-api-access-vtzqm\") pod \"7458cfeb-9419-45cb-adf9-05c35f988b3e\" (UID: \"7458cfeb-9419-45cb-adf9-05c35f988b3e\") " Dec 06 16:23:26 crc kubenswrapper[4813]: I1206 16:23:26.354700 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7458cfeb-9419-45cb-adf9-05c35f988b3e-utilities\") pod \"7458cfeb-9419-45cb-adf9-05c35f988b3e\" (UID: \"7458cfeb-9419-45cb-adf9-05c35f988b3e\") " Dec 06 16:23:26 crc kubenswrapper[4813]: I1206 16:23:26.356558 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7458cfeb-9419-45cb-adf9-05c35f988b3e-utilities" (OuterVolumeSpecName: "utilities") pod "7458cfeb-9419-45cb-adf9-05c35f988b3e" (UID: "7458cfeb-9419-45cb-adf9-05c35f988b3e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:23:26 crc kubenswrapper[4813]: I1206 16:23:26.365516 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7458cfeb-9419-45cb-adf9-05c35f988b3e-kube-api-access-vtzqm" (OuterVolumeSpecName: "kube-api-access-vtzqm") pod "7458cfeb-9419-45cb-adf9-05c35f988b3e" (UID: "7458cfeb-9419-45cb-adf9-05c35f988b3e"). InnerVolumeSpecName "kube-api-access-vtzqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:23:26 crc kubenswrapper[4813]: I1206 16:23:26.456688 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtzqm\" (UniqueName: \"kubernetes.io/projected/7458cfeb-9419-45cb-adf9-05c35f988b3e-kube-api-access-vtzqm\") on node \"crc\" DevicePath \"\"" Dec 06 16:23:26 crc kubenswrapper[4813]: I1206 16:23:26.456715 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7458cfeb-9419-45cb-adf9-05c35f988b3e-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 16:23:26 crc kubenswrapper[4813]: I1206 16:23:26.469557 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7458cfeb-9419-45cb-adf9-05c35f988b3e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7458cfeb-9419-45cb-adf9-05c35f988b3e" (UID: "7458cfeb-9419-45cb-adf9-05c35f988b3e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:23:26 crc kubenswrapper[4813]: I1206 16:23:26.558081 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7458cfeb-9419-45cb-adf9-05c35f988b3e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 16:23:26 crc kubenswrapper[4813]: I1206 16:23:26.746493 4813 generic.go:334] "Generic (PLEG): container finished" podID="7458cfeb-9419-45cb-adf9-05c35f988b3e" containerID="a7deb66f46515059fc25463214cea26f2d7f26e7b32559ac73387eea17ae7159" exitCode=0 Dec 06 16:23:26 crc kubenswrapper[4813]: I1206 16:23:26.746554 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bqsvv" event={"ID":"7458cfeb-9419-45cb-adf9-05c35f988b3e","Type":"ContainerDied","Data":"a7deb66f46515059fc25463214cea26f2d7f26e7b32559ac73387eea17ae7159"} Dec 06 16:23:26 crc kubenswrapper[4813]: I1206 16:23:26.746594 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bqsvv" event={"ID":"7458cfeb-9419-45cb-adf9-05c35f988b3e","Type":"ContainerDied","Data":"71db26c103397ddcd25c33fceacaa0d91436c852aff7794316fbb4539e1f5791"} Dec 06 16:23:26 crc kubenswrapper[4813]: I1206 16:23:26.746622 4813 scope.go:117] "RemoveContainer" containerID="a7deb66f46515059fc25463214cea26f2d7f26e7b32559ac73387eea17ae7159" Dec 06 16:23:26 crc kubenswrapper[4813]: I1206 16:23:26.746802 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bqsvv" Dec 06 16:23:26 crc kubenswrapper[4813]: I1206 16:23:26.783404 4813 scope.go:117] "RemoveContainer" containerID="39faa1f9faeadcb062a911288bac9f85c5f5515527c46edb59329a12c8224f17" Dec 06 16:23:26 crc kubenswrapper[4813]: I1206 16:23:26.786287 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bqsvv"] Dec 06 16:23:26 crc kubenswrapper[4813]: I1206 16:23:26.799134 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bqsvv"] Dec 06 16:23:26 crc kubenswrapper[4813]: I1206 16:23:26.811192 4813 scope.go:117] "RemoveContainer" containerID="72f4f99fa784c6935864342b4e9a1f216b6626762f4dd51057a8ea202ad7f645" Dec 06 16:23:26 crc kubenswrapper[4813]: I1206 16:23:26.871807 4813 scope.go:117] "RemoveContainer" containerID="a7deb66f46515059fc25463214cea26f2d7f26e7b32559ac73387eea17ae7159" Dec 06 16:23:26 crc kubenswrapper[4813]: E1206 16:23:26.872421 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7deb66f46515059fc25463214cea26f2d7f26e7b32559ac73387eea17ae7159\": container with ID starting with a7deb66f46515059fc25463214cea26f2d7f26e7b32559ac73387eea17ae7159 not found: ID does not exist" containerID="a7deb66f46515059fc25463214cea26f2d7f26e7b32559ac73387eea17ae7159" Dec 06 16:23:26 crc kubenswrapper[4813]: I1206 16:23:26.872476 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7deb66f46515059fc25463214cea26f2d7f26e7b32559ac73387eea17ae7159"} err="failed to get container status \"a7deb66f46515059fc25463214cea26f2d7f26e7b32559ac73387eea17ae7159\": rpc error: code = NotFound desc = could not find container \"a7deb66f46515059fc25463214cea26f2d7f26e7b32559ac73387eea17ae7159\": container with ID starting with a7deb66f46515059fc25463214cea26f2d7f26e7b32559ac73387eea17ae7159 not found: ID does not exist" Dec 06 16:23:26 crc kubenswrapper[4813]: I1206 16:23:26.872513 4813 scope.go:117] "RemoveContainer" containerID="39faa1f9faeadcb062a911288bac9f85c5f5515527c46edb59329a12c8224f17" Dec 06 16:23:26 crc kubenswrapper[4813]: E1206 16:23:26.872917 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39faa1f9faeadcb062a911288bac9f85c5f5515527c46edb59329a12c8224f17\": container with ID starting with 39faa1f9faeadcb062a911288bac9f85c5f5515527c46edb59329a12c8224f17 not found: ID does not exist" containerID="39faa1f9faeadcb062a911288bac9f85c5f5515527c46edb59329a12c8224f17" Dec 06 16:23:26 crc kubenswrapper[4813]: I1206 16:23:26.872951 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39faa1f9faeadcb062a911288bac9f85c5f5515527c46edb59329a12c8224f17"} err="failed to get container status \"39faa1f9faeadcb062a911288bac9f85c5f5515527c46edb59329a12c8224f17\": rpc error: code = NotFound desc = could not find container \"39faa1f9faeadcb062a911288bac9f85c5f5515527c46edb59329a12c8224f17\": container with ID starting with 39faa1f9faeadcb062a911288bac9f85c5f5515527c46edb59329a12c8224f17 not found: ID does not exist" Dec 06 16:23:26 crc kubenswrapper[4813]: I1206 16:23:26.873044 4813 scope.go:117] "RemoveContainer" containerID="72f4f99fa784c6935864342b4e9a1f216b6626762f4dd51057a8ea202ad7f645" Dec 06 16:23:26 crc kubenswrapper[4813]: E1206 16:23:26.873426 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72f4f99fa784c6935864342b4e9a1f216b6626762f4dd51057a8ea202ad7f645\": container with ID starting with 72f4f99fa784c6935864342b4e9a1f216b6626762f4dd51057a8ea202ad7f645 not found: ID does not exist" containerID="72f4f99fa784c6935864342b4e9a1f216b6626762f4dd51057a8ea202ad7f645" Dec 06 16:23:26 crc kubenswrapper[4813]: I1206 16:23:26.873470 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72f4f99fa784c6935864342b4e9a1f216b6626762f4dd51057a8ea202ad7f645"} err="failed to get container status \"72f4f99fa784c6935864342b4e9a1f216b6626762f4dd51057a8ea202ad7f645\": rpc error: code = NotFound desc = could not find container \"72f4f99fa784c6935864342b4e9a1f216b6626762f4dd51057a8ea202ad7f645\": container with ID starting with 72f4f99fa784c6935864342b4e9a1f216b6626762f4dd51057a8ea202ad7f645 not found: ID does not exist" Dec 06 16:23:28 crc kubenswrapper[4813]: I1206 16:23:28.502122 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7458cfeb-9419-45cb-adf9-05c35f988b3e" path="/var/lib/kubelet/pods/7458cfeb-9419-45cb-adf9-05c35f988b3e/volumes" Dec 06 16:23:49 crc kubenswrapper[4813]: I1206 16:23:49.427841 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:23:49 crc kubenswrapper[4813]: I1206 16:23:49.429493 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:23:49 crc kubenswrapper[4813]: I1206 16:23:49.429662 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" Dec 06 16:23:49 crc kubenswrapper[4813]: I1206 16:23:49.430503 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1b8cc98443ea491aa620632d54d48664f2e8cd65b0790f1013b24051a06625ca"} pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 16:23:49 crc kubenswrapper[4813]: I1206 16:23:49.430656 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" containerID="cri-o://1b8cc98443ea491aa620632d54d48664f2e8cd65b0790f1013b24051a06625ca" gracePeriod=600 Dec 06 16:23:49 crc kubenswrapper[4813]: E1206 16:23:49.550863 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:23:49 crc kubenswrapper[4813]: I1206 16:23:49.983933 4813 generic.go:334] "Generic (PLEG): container finished" podID="d88e8bae-c055-4c55-b548-f621ff96de06" containerID="1b8cc98443ea491aa620632d54d48664f2e8cd65b0790f1013b24051a06625ca" exitCode=0 Dec 06 16:23:49 crc kubenswrapper[4813]: I1206 16:23:49.983963 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerDied","Data":"1b8cc98443ea491aa620632d54d48664f2e8cd65b0790f1013b24051a06625ca"} Dec 06 16:23:49 crc kubenswrapper[4813]: I1206 16:23:49.984041 4813 scope.go:117] "RemoveContainer" containerID="f475b8b00603d232fa96fab484dd38d82f5207920b09491a14705e92bb65713d" Dec 06 16:23:49 crc kubenswrapper[4813]: I1206 16:23:49.985471 4813 scope.go:117] "RemoveContainer" containerID="1b8cc98443ea491aa620632d54d48664f2e8cd65b0790f1013b24051a06625ca" Dec 06 16:23:49 crc kubenswrapper[4813]: E1206 16:23:49.986098 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:24:04 crc kubenswrapper[4813]: I1206 16:24:04.488232 4813 scope.go:117] "RemoveContainer" containerID="1b8cc98443ea491aa620632d54d48664f2e8cd65b0790f1013b24051a06625ca" Dec 06 16:24:04 crc kubenswrapper[4813]: E1206 16:24:04.489432 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:24:15 crc kubenswrapper[4813]: I1206 16:24:15.488173 4813 scope.go:117] "RemoveContainer" containerID="1b8cc98443ea491aa620632d54d48664f2e8cd65b0790f1013b24051a06625ca" Dec 06 16:24:15 crc kubenswrapper[4813]: E1206 16:24:15.488922 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:24:28 crc kubenswrapper[4813]: I1206 16:24:28.487504 4813 scope.go:117] "RemoveContainer" containerID="1b8cc98443ea491aa620632d54d48664f2e8cd65b0790f1013b24051a06625ca" Dec 06 16:24:28 crc kubenswrapper[4813]: E1206 16:24:28.488224 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:24:43 crc kubenswrapper[4813]: I1206 16:24:43.488063 4813 scope.go:117] "RemoveContainer" containerID="1b8cc98443ea491aa620632d54d48664f2e8cd65b0790f1013b24051a06625ca" Dec 06 16:24:43 crc kubenswrapper[4813]: E1206 16:24:43.488911 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:24:56 crc kubenswrapper[4813]: I1206 16:24:56.487610 4813 scope.go:117] "RemoveContainer" containerID="1b8cc98443ea491aa620632d54d48664f2e8cd65b0790f1013b24051a06625ca" Dec 06 16:24:56 crc kubenswrapper[4813]: E1206 16:24:56.490311 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:25:07 crc kubenswrapper[4813]: I1206 16:25:07.486780 4813 scope.go:117] "RemoveContainer" containerID="1b8cc98443ea491aa620632d54d48664f2e8cd65b0790f1013b24051a06625ca" Dec 06 16:25:07 crc kubenswrapper[4813]: E1206 16:25:07.488634 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:25:19 crc kubenswrapper[4813]: I1206 16:25:19.487393 4813 scope.go:117] "RemoveContainer" containerID="1b8cc98443ea491aa620632d54d48664f2e8cd65b0790f1013b24051a06625ca" Dec 06 16:25:19 crc kubenswrapper[4813]: E1206 16:25:19.488314 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:25:22 crc kubenswrapper[4813]: I1206 16:25:22.920618 4813 generic.go:334] "Generic (PLEG): container finished" podID="bfa11078-d153-43a2-abc4-b70f491dc4a6" containerID="814ae62233afc9678a4c073b9a4fec1cf6f15e78b47d1bfa35830574e9f5a35b" exitCode=0 Dec 06 16:25:22 crc kubenswrapper[4813]: I1206 16:25:22.920717 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j" event={"ID":"bfa11078-d153-43a2-abc4-b70f491dc4a6","Type":"ContainerDied","Data":"814ae62233afc9678a4c073b9a4fec1cf6f15e78b47d1bfa35830574e9f5a35b"} Dec 06 16:25:24 crc kubenswrapper[4813]: I1206 16:25:24.379209 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j" Dec 06 16:25:24 crc kubenswrapper[4813]: I1206 16:25:24.542798 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d628l\" (UniqueName: \"kubernetes.io/projected/bfa11078-d153-43a2-abc4-b70f491dc4a6-kube-api-access-d628l\") pod \"bfa11078-d153-43a2-abc4-b70f491dc4a6\" (UID: \"bfa11078-d153-43a2-abc4-b70f491dc4a6\") " Dec 06 16:25:24 crc kubenswrapper[4813]: I1206 16:25:24.542979 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfa11078-d153-43a2-abc4-b70f491dc4a6-libvirt-combined-ca-bundle\") pod \"bfa11078-d153-43a2-abc4-b70f491dc4a6\" (UID: \"bfa11078-d153-43a2-abc4-b70f491dc4a6\") " Dec 06 16:25:24 crc kubenswrapper[4813]: I1206 16:25:24.543098 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/bfa11078-d153-43a2-abc4-b70f491dc4a6-libvirt-secret-0\") pod \"bfa11078-d153-43a2-abc4-b70f491dc4a6\" (UID: \"bfa11078-d153-43a2-abc4-b70f491dc4a6\") " Dec 06 16:25:24 crc kubenswrapper[4813]: I1206 16:25:24.543127 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bfa11078-d153-43a2-abc4-b70f491dc4a6-ssh-key\") pod \"bfa11078-d153-43a2-abc4-b70f491dc4a6\" (UID: \"bfa11078-d153-43a2-abc4-b70f491dc4a6\") " Dec 06 16:25:24 crc kubenswrapper[4813]: I1206 16:25:24.543170 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bfa11078-d153-43a2-abc4-b70f491dc4a6-inventory\") pod \"bfa11078-d153-43a2-abc4-b70f491dc4a6\" (UID: \"bfa11078-d153-43a2-abc4-b70f491dc4a6\") " Dec 06 16:25:24 crc kubenswrapper[4813]: I1206 16:25:24.549866 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfa11078-d153-43a2-abc4-b70f491dc4a6-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "bfa11078-d153-43a2-abc4-b70f491dc4a6" (UID: "bfa11078-d153-43a2-abc4-b70f491dc4a6"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:25:24 crc kubenswrapper[4813]: I1206 16:25:24.553924 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfa11078-d153-43a2-abc4-b70f491dc4a6-kube-api-access-d628l" (OuterVolumeSpecName: "kube-api-access-d628l") pod "bfa11078-d153-43a2-abc4-b70f491dc4a6" (UID: "bfa11078-d153-43a2-abc4-b70f491dc4a6"). InnerVolumeSpecName "kube-api-access-d628l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:25:24 crc kubenswrapper[4813]: I1206 16:25:24.577036 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfa11078-d153-43a2-abc4-b70f491dc4a6-inventory" (OuterVolumeSpecName: "inventory") pod "bfa11078-d153-43a2-abc4-b70f491dc4a6" (UID: "bfa11078-d153-43a2-abc4-b70f491dc4a6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:25:24 crc kubenswrapper[4813]: I1206 16:25:24.585133 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfa11078-d153-43a2-abc4-b70f491dc4a6-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "bfa11078-d153-43a2-abc4-b70f491dc4a6" (UID: "bfa11078-d153-43a2-abc4-b70f491dc4a6"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:25:24 crc kubenswrapper[4813]: I1206 16:25:24.592205 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfa11078-d153-43a2-abc4-b70f491dc4a6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bfa11078-d153-43a2-abc4-b70f491dc4a6" (UID: "bfa11078-d153-43a2-abc4-b70f491dc4a6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:25:24 crc kubenswrapper[4813]: I1206 16:25:24.647129 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d628l\" (UniqueName: \"kubernetes.io/projected/bfa11078-d153-43a2-abc4-b70f491dc4a6-kube-api-access-d628l\") on node \"crc\" DevicePath \"\"" Dec 06 16:25:24 crc kubenswrapper[4813]: I1206 16:25:24.647165 4813 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfa11078-d153-43a2-abc4-b70f491dc4a6-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:25:24 crc kubenswrapper[4813]: I1206 16:25:24.647179 4813 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/bfa11078-d153-43a2-abc4-b70f491dc4a6-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 06 16:25:24 crc kubenswrapper[4813]: I1206 16:25:24.647191 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bfa11078-d153-43a2-abc4-b70f491dc4a6-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 16:25:24 crc kubenswrapper[4813]: I1206 16:25:24.647203 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bfa11078-d153-43a2-abc4-b70f491dc4a6-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 16:25:24 crc kubenswrapper[4813]: I1206 16:25:24.950670 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j" event={"ID":"bfa11078-d153-43a2-abc4-b70f491dc4a6","Type":"ContainerDied","Data":"d5914d41175b97c87ad4c0cedfe6473ab888f5a698b85a8b551c385d77dd5efd"} Dec 06 16:25:24 crc kubenswrapper[4813]: I1206 16:25:24.951140 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5914d41175b97c87ad4c0cedfe6473ab888f5a698b85a8b551c385d77dd5efd" Dec 06 16:25:24 crc kubenswrapper[4813]: I1206 16:25:24.950807 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.061655 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s"] Dec 06 16:25:25 crc kubenswrapper[4813]: E1206 16:25:25.062449 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfa11078-d153-43a2-abc4-b70f491dc4a6" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.062576 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfa11078-d153-43a2-abc4-b70f491dc4a6" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 06 16:25:25 crc kubenswrapper[4813]: E1206 16:25:25.062660 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7458cfeb-9419-45cb-adf9-05c35f988b3e" containerName="registry-server" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.062732 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="7458cfeb-9419-45cb-adf9-05c35f988b3e" containerName="registry-server" Dec 06 16:25:25 crc kubenswrapper[4813]: E1206 16:25:25.062843 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7458cfeb-9419-45cb-adf9-05c35f988b3e" containerName="extract-content" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.062923 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="7458cfeb-9419-45cb-adf9-05c35f988b3e" containerName="extract-content" Dec 06 16:25:25 crc kubenswrapper[4813]: E1206 16:25:25.063005 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7458cfeb-9419-45cb-adf9-05c35f988b3e" containerName="extract-utilities" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.063073 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="7458cfeb-9419-45cb-adf9-05c35f988b3e" containerName="extract-utilities" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.063419 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfa11078-d153-43a2-abc4-b70f491dc4a6" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.063546 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="7458cfeb-9419-45cb-adf9-05c35f988b3e" containerName="registry-server" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.064471 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.070692 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s"] Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.071821 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.071968 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.072545 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.072857 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jvkhf" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.073011 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.073093 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.073223 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.258170 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mfh5s\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.258244 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mfh5s\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.258332 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7ba82e65-8192-4943-8a2a-863d041ad231-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mfh5s\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.258364 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mfh5s\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.258386 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mfh5s\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.258433 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mfh5s\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.258460 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mfh5s\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.258491 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mfh5s\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.258547 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mz72t\" (UniqueName: \"kubernetes.io/projected/7ba82e65-8192-4943-8a2a-863d041ad231-kube-api-access-mz72t\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mfh5s\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.360341 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7ba82e65-8192-4943-8a2a-863d041ad231-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mfh5s\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.360394 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mfh5s\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.360425 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mfh5s\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.360475 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mfh5s\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.360506 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mfh5s\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.360538 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mfh5s\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.360589 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mz72t\" (UniqueName: \"kubernetes.io/projected/7ba82e65-8192-4943-8a2a-863d041ad231-kube-api-access-mz72t\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mfh5s\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.360655 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mfh5s\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.360706 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mfh5s\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.365964 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mfh5s\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.366328 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mfh5s\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.366519 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7ba82e65-8192-4943-8a2a-863d041ad231-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mfh5s\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.367420 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mfh5s\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.368404 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mfh5s\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.370230 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mfh5s\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.370513 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mfh5s\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.431404 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mfh5s\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.433727 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mz72t\" (UniqueName: \"kubernetes.io/projected/7ba82e65-8192-4943-8a2a-863d041ad231-kube-api-access-mz72t\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mfh5s\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:25:25 crc kubenswrapper[4813]: I1206 16:25:25.679327 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:25:26 crc kubenswrapper[4813]: I1206 16:25:26.223323 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s"] Dec 06 16:25:26 crc kubenswrapper[4813]: I1206 16:25:26.985866 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" event={"ID":"7ba82e65-8192-4943-8a2a-863d041ad231","Type":"ContainerStarted","Data":"4550799567e52d98805e1acc6be838dcbf81b8c05a15852f901b87bd9e20da23"} Dec 06 16:25:28 crc kubenswrapper[4813]: I1206 16:25:28.001075 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" event={"ID":"7ba82e65-8192-4943-8a2a-863d041ad231","Type":"ContainerStarted","Data":"a68abb34910726f382e7410aabbe8b2851f441e06609b7ea07f5e963ca5195e8"} Dec 06 16:25:28 crc kubenswrapper[4813]: I1206 16:25:28.045009 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" podStartSLOduration=2.376514144 podStartE2EDuration="3.044987004s" podCreationTimestamp="2025-12-06 16:25:25 +0000 UTC" firstStartedPulling="2025-12-06 16:25:26.230178551 +0000 UTC m=+2366.121058137" lastFinishedPulling="2025-12-06 16:25:26.898651411 +0000 UTC m=+2366.789530997" observedRunningTime="2025-12-06 16:25:28.035706219 +0000 UTC m=+2367.926585805" watchObservedRunningTime="2025-12-06 16:25:28.044987004 +0000 UTC m=+2367.935866600" Dec 06 16:25:30 crc kubenswrapper[4813]: I1206 16:25:30.491988 4813 scope.go:117] "RemoveContainer" containerID="1b8cc98443ea491aa620632d54d48664f2e8cd65b0790f1013b24051a06625ca" Dec 06 16:25:30 crc kubenswrapper[4813]: E1206 16:25:30.492428 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:25:43 crc kubenswrapper[4813]: I1206 16:25:43.487188 4813 scope.go:117] "RemoveContainer" containerID="1b8cc98443ea491aa620632d54d48664f2e8cd65b0790f1013b24051a06625ca" Dec 06 16:25:43 crc kubenswrapper[4813]: E1206 16:25:43.487962 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:25:58 crc kubenswrapper[4813]: I1206 16:25:58.487472 4813 scope.go:117] "RemoveContainer" containerID="1b8cc98443ea491aa620632d54d48664f2e8cd65b0790f1013b24051a06625ca" Dec 06 16:25:58 crc kubenswrapper[4813]: E1206 16:25:58.488237 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:26:11 crc kubenswrapper[4813]: I1206 16:26:11.488796 4813 scope.go:117] "RemoveContainer" containerID="1b8cc98443ea491aa620632d54d48664f2e8cd65b0790f1013b24051a06625ca" Dec 06 16:26:11 crc kubenswrapper[4813]: E1206 16:26:11.490342 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:26:26 crc kubenswrapper[4813]: I1206 16:26:26.486606 4813 scope.go:117] "RemoveContainer" containerID="1b8cc98443ea491aa620632d54d48664f2e8cd65b0790f1013b24051a06625ca" Dec 06 16:26:26 crc kubenswrapper[4813]: E1206 16:26:26.487379 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:26:38 crc kubenswrapper[4813]: I1206 16:26:38.486535 4813 scope.go:117] "RemoveContainer" containerID="1b8cc98443ea491aa620632d54d48664f2e8cd65b0790f1013b24051a06625ca" Dec 06 16:26:38 crc kubenswrapper[4813]: E1206 16:26:38.487213 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:26:50 crc kubenswrapper[4813]: I1206 16:26:50.493229 4813 scope.go:117] "RemoveContainer" containerID="1b8cc98443ea491aa620632d54d48664f2e8cd65b0790f1013b24051a06625ca" Dec 06 16:26:50 crc kubenswrapper[4813]: E1206 16:26:50.493823 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:26:57 crc kubenswrapper[4813]: I1206 16:26:57.200481 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bt86c"] Dec 06 16:26:57 crc kubenswrapper[4813]: I1206 16:26:57.204307 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bt86c" Dec 06 16:26:57 crc kubenswrapper[4813]: I1206 16:26:57.225158 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bt86c"] Dec 06 16:26:57 crc kubenswrapper[4813]: I1206 16:26:57.381983 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d049dac5-fc23-4b86-ab68-c114ed0b6371-utilities\") pod \"certified-operators-bt86c\" (UID: \"d049dac5-fc23-4b86-ab68-c114ed0b6371\") " pod="openshift-marketplace/certified-operators-bt86c" Dec 06 16:26:57 crc kubenswrapper[4813]: I1206 16:26:57.382147 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d049dac5-fc23-4b86-ab68-c114ed0b6371-catalog-content\") pod \"certified-operators-bt86c\" (UID: \"d049dac5-fc23-4b86-ab68-c114ed0b6371\") " pod="openshift-marketplace/certified-operators-bt86c" Dec 06 16:26:57 crc kubenswrapper[4813]: I1206 16:26:57.382194 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bkhs\" (UniqueName: \"kubernetes.io/projected/d049dac5-fc23-4b86-ab68-c114ed0b6371-kube-api-access-8bkhs\") pod \"certified-operators-bt86c\" (UID: \"d049dac5-fc23-4b86-ab68-c114ed0b6371\") " pod="openshift-marketplace/certified-operators-bt86c" Dec 06 16:26:57 crc kubenswrapper[4813]: I1206 16:26:57.484247 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d049dac5-fc23-4b86-ab68-c114ed0b6371-utilities\") pod \"certified-operators-bt86c\" (UID: \"d049dac5-fc23-4b86-ab68-c114ed0b6371\") " pod="openshift-marketplace/certified-operators-bt86c" Dec 06 16:26:57 crc kubenswrapper[4813]: I1206 16:26:57.484399 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d049dac5-fc23-4b86-ab68-c114ed0b6371-catalog-content\") pod \"certified-operators-bt86c\" (UID: \"d049dac5-fc23-4b86-ab68-c114ed0b6371\") " pod="openshift-marketplace/certified-operators-bt86c" Dec 06 16:26:57 crc kubenswrapper[4813]: I1206 16:26:57.484425 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bkhs\" (UniqueName: \"kubernetes.io/projected/d049dac5-fc23-4b86-ab68-c114ed0b6371-kube-api-access-8bkhs\") pod \"certified-operators-bt86c\" (UID: \"d049dac5-fc23-4b86-ab68-c114ed0b6371\") " pod="openshift-marketplace/certified-operators-bt86c" Dec 06 16:26:57 crc kubenswrapper[4813]: I1206 16:26:57.484797 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d049dac5-fc23-4b86-ab68-c114ed0b6371-utilities\") pod \"certified-operators-bt86c\" (UID: \"d049dac5-fc23-4b86-ab68-c114ed0b6371\") " pod="openshift-marketplace/certified-operators-bt86c" Dec 06 16:26:57 crc kubenswrapper[4813]: I1206 16:26:57.484943 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d049dac5-fc23-4b86-ab68-c114ed0b6371-catalog-content\") pod \"certified-operators-bt86c\" (UID: \"d049dac5-fc23-4b86-ab68-c114ed0b6371\") " pod="openshift-marketplace/certified-operators-bt86c" Dec 06 16:26:57 crc kubenswrapper[4813]: I1206 16:26:57.503646 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bkhs\" (UniqueName: \"kubernetes.io/projected/d049dac5-fc23-4b86-ab68-c114ed0b6371-kube-api-access-8bkhs\") pod \"certified-operators-bt86c\" (UID: \"d049dac5-fc23-4b86-ab68-c114ed0b6371\") " pod="openshift-marketplace/certified-operators-bt86c" Dec 06 16:26:57 crc kubenswrapper[4813]: I1206 16:26:57.589386 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bt86c" Dec 06 16:26:58 crc kubenswrapper[4813]: I1206 16:26:58.097901 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bt86c"] Dec 06 16:26:59 crc kubenswrapper[4813]: I1206 16:26:59.018866 4813 generic.go:334] "Generic (PLEG): container finished" podID="d049dac5-fc23-4b86-ab68-c114ed0b6371" containerID="9df454d2b2605163125c03671cb312630b20cdac92eef3791d37b1317bb2af76" exitCode=0 Dec 06 16:26:59 crc kubenswrapper[4813]: I1206 16:26:59.019045 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bt86c" event={"ID":"d049dac5-fc23-4b86-ab68-c114ed0b6371","Type":"ContainerDied","Data":"9df454d2b2605163125c03671cb312630b20cdac92eef3791d37b1317bb2af76"} Dec 06 16:26:59 crc kubenswrapper[4813]: I1206 16:26:59.019220 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bt86c" event={"ID":"d049dac5-fc23-4b86-ab68-c114ed0b6371","Type":"ContainerStarted","Data":"70c7dc0ac96b37def8d4e6db71511781926188693d9f239b6d466accbaeb2c49"} Dec 06 16:27:01 crc kubenswrapper[4813]: I1206 16:27:01.039125 4813 generic.go:334] "Generic (PLEG): container finished" podID="d049dac5-fc23-4b86-ab68-c114ed0b6371" containerID="522692386d728b8dce9d5bb112e19798cd1d98d5b857c0e28ad11d7c86cf4ef9" exitCode=0 Dec 06 16:27:01 crc kubenswrapper[4813]: I1206 16:27:01.039188 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bt86c" event={"ID":"d049dac5-fc23-4b86-ab68-c114ed0b6371","Type":"ContainerDied","Data":"522692386d728b8dce9d5bb112e19798cd1d98d5b857c0e28ad11d7c86cf4ef9"} Dec 06 16:27:02 crc kubenswrapper[4813]: I1206 16:27:02.054370 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bt86c" event={"ID":"d049dac5-fc23-4b86-ab68-c114ed0b6371","Type":"ContainerStarted","Data":"1533c09e18f21d9a780882d5d66a33815f868c89101b2116825d0d872209c1bc"} Dec 06 16:27:02 crc kubenswrapper[4813]: I1206 16:27:02.096972 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bt86c" podStartSLOduration=2.654222733 podStartE2EDuration="5.096951173s" podCreationTimestamp="2025-12-06 16:26:57 +0000 UTC" firstStartedPulling="2025-12-06 16:26:59.020957203 +0000 UTC m=+2458.911836819" lastFinishedPulling="2025-12-06 16:27:01.463685643 +0000 UTC m=+2461.354565259" observedRunningTime="2025-12-06 16:27:02.083511348 +0000 UTC m=+2461.974390964" watchObservedRunningTime="2025-12-06 16:27:02.096951173 +0000 UTC m=+2461.987830759" Dec 06 16:27:02 crc kubenswrapper[4813]: I1206 16:27:02.487618 4813 scope.go:117] "RemoveContainer" containerID="1b8cc98443ea491aa620632d54d48664f2e8cd65b0790f1013b24051a06625ca" Dec 06 16:27:02 crc kubenswrapper[4813]: E1206 16:27:02.487911 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:27:07 crc kubenswrapper[4813]: I1206 16:27:07.590974 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bt86c" Dec 06 16:27:07 crc kubenswrapper[4813]: I1206 16:27:07.591726 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bt86c" Dec 06 16:27:07 crc kubenswrapper[4813]: I1206 16:27:07.674063 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bt86c" Dec 06 16:27:08 crc kubenswrapper[4813]: I1206 16:27:08.195011 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bt86c" Dec 06 16:27:08 crc kubenswrapper[4813]: I1206 16:27:08.288661 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bt86c"] Dec 06 16:27:10 crc kubenswrapper[4813]: I1206 16:27:10.139101 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bt86c" podUID="d049dac5-fc23-4b86-ab68-c114ed0b6371" containerName="registry-server" containerID="cri-o://1533c09e18f21d9a780882d5d66a33815f868c89101b2116825d0d872209c1bc" gracePeriod=2 Dec 06 16:27:10 crc kubenswrapper[4813]: I1206 16:27:10.683391 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bt86c" Dec 06 16:27:10 crc kubenswrapper[4813]: I1206 16:27:10.778666 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d049dac5-fc23-4b86-ab68-c114ed0b6371-utilities\") pod \"d049dac5-fc23-4b86-ab68-c114ed0b6371\" (UID: \"d049dac5-fc23-4b86-ab68-c114ed0b6371\") " Dec 06 16:27:10 crc kubenswrapper[4813]: I1206 16:27:10.779591 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d049dac5-fc23-4b86-ab68-c114ed0b6371-utilities" (OuterVolumeSpecName: "utilities") pod "d049dac5-fc23-4b86-ab68-c114ed0b6371" (UID: "d049dac5-fc23-4b86-ab68-c114ed0b6371"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:27:10 crc kubenswrapper[4813]: I1206 16:27:10.880866 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bkhs\" (UniqueName: \"kubernetes.io/projected/d049dac5-fc23-4b86-ab68-c114ed0b6371-kube-api-access-8bkhs\") pod \"d049dac5-fc23-4b86-ab68-c114ed0b6371\" (UID: \"d049dac5-fc23-4b86-ab68-c114ed0b6371\") " Dec 06 16:27:10 crc kubenswrapper[4813]: I1206 16:27:10.881155 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d049dac5-fc23-4b86-ab68-c114ed0b6371-catalog-content\") pod \"d049dac5-fc23-4b86-ab68-c114ed0b6371\" (UID: \"d049dac5-fc23-4b86-ab68-c114ed0b6371\") " Dec 06 16:27:10 crc kubenswrapper[4813]: I1206 16:27:10.881624 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d049dac5-fc23-4b86-ab68-c114ed0b6371-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 16:27:10 crc kubenswrapper[4813]: I1206 16:27:10.886292 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d049dac5-fc23-4b86-ab68-c114ed0b6371-kube-api-access-8bkhs" (OuterVolumeSpecName: "kube-api-access-8bkhs") pod "d049dac5-fc23-4b86-ab68-c114ed0b6371" (UID: "d049dac5-fc23-4b86-ab68-c114ed0b6371"). InnerVolumeSpecName "kube-api-access-8bkhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:27:10 crc kubenswrapper[4813]: I1206 16:27:10.926494 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d049dac5-fc23-4b86-ab68-c114ed0b6371-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d049dac5-fc23-4b86-ab68-c114ed0b6371" (UID: "d049dac5-fc23-4b86-ab68-c114ed0b6371"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:27:10 crc kubenswrapper[4813]: I1206 16:27:10.982560 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d049dac5-fc23-4b86-ab68-c114ed0b6371-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 16:27:10 crc kubenswrapper[4813]: I1206 16:27:10.982593 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bkhs\" (UniqueName: \"kubernetes.io/projected/d049dac5-fc23-4b86-ab68-c114ed0b6371-kube-api-access-8bkhs\") on node \"crc\" DevicePath \"\"" Dec 06 16:27:11 crc kubenswrapper[4813]: I1206 16:27:11.151538 4813 generic.go:334] "Generic (PLEG): container finished" podID="d049dac5-fc23-4b86-ab68-c114ed0b6371" containerID="1533c09e18f21d9a780882d5d66a33815f868c89101b2116825d0d872209c1bc" exitCode=0 Dec 06 16:27:11 crc kubenswrapper[4813]: I1206 16:27:11.151591 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bt86c" event={"ID":"d049dac5-fc23-4b86-ab68-c114ed0b6371","Type":"ContainerDied","Data":"1533c09e18f21d9a780882d5d66a33815f868c89101b2116825d0d872209c1bc"} Dec 06 16:27:11 crc kubenswrapper[4813]: I1206 16:27:11.151639 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bt86c" event={"ID":"d049dac5-fc23-4b86-ab68-c114ed0b6371","Type":"ContainerDied","Data":"70c7dc0ac96b37def8d4e6db71511781926188693d9f239b6d466accbaeb2c49"} Dec 06 16:27:11 crc kubenswrapper[4813]: I1206 16:27:11.151668 4813 scope.go:117] "RemoveContainer" containerID="1533c09e18f21d9a780882d5d66a33815f868c89101b2116825d0d872209c1bc" Dec 06 16:27:11 crc kubenswrapper[4813]: I1206 16:27:11.153009 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bt86c" Dec 06 16:27:11 crc kubenswrapper[4813]: I1206 16:27:11.175119 4813 scope.go:117] "RemoveContainer" containerID="522692386d728b8dce9d5bb112e19798cd1d98d5b857c0e28ad11d7c86cf4ef9" Dec 06 16:27:11 crc kubenswrapper[4813]: I1206 16:27:11.221752 4813 scope.go:117] "RemoveContainer" containerID="9df454d2b2605163125c03671cb312630b20cdac92eef3791d37b1317bb2af76" Dec 06 16:27:11 crc kubenswrapper[4813]: I1206 16:27:11.221879 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bt86c"] Dec 06 16:27:11 crc kubenswrapper[4813]: I1206 16:27:11.233436 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bt86c"] Dec 06 16:27:11 crc kubenswrapper[4813]: I1206 16:27:11.267197 4813 scope.go:117] "RemoveContainer" containerID="1533c09e18f21d9a780882d5d66a33815f868c89101b2116825d0d872209c1bc" Dec 06 16:27:11 crc kubenswrapper[4813]: E1206 16:27:11.267848 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1533c09e18f21d9a780882d5d66a33815f868c89101b2116825d0d872209c1bc\": container with ID starting with 1533c09e18f21d9a780882d5d66a33815f868c89101b2116825d0d872209c1bc not found: ID does not exist" containerID="1533c09e18f21d9a780882d5d66a33815f868c89101b2116825d0d872209c1bc" Dec 06 16:27:11 crc kubenswrapper[4813]: I1206 16:27:11.267900 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1533c09e18f21d9a780882d5d66a33815f868c89101b2116825d0d872209c1bc"} err="failed to get container status \"1533c09e18f21d9a780882d5d66a33815f868c89101b2116825d0d872209c1bc\": rpc error: code = NotFound desc = could not find container \"1533c09e18f21d9a780882d5d66a33815f868c89101b2116825d0d872209c1bc\": container with ID starting with 1533c09e18f21d9a780882d5d66a33815f868c89101b2116825d0d872209c1bc not found: ID does not exist" Dec 06 16:27:11 crc kubenswrapper[4813]: I1206 16:27:11.267933 4813 scope.go:117] "RemoveContainer" containerID="522692386d728b8dce9d5bb112e19798cd1d98d5b857c0e28ad11d7c86cf4ef9" Dec 06 16:27:11 crc kubenswrapper[4813]: E1206 16:27:11.268706 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"522692386d728b8dce9d5bb112e19798cd1d98d5b857c0e28ad11d7c86cf4ef9\": container with ID starting with 522692386d728b8dce9d5bb112e19798cd1d98d5b857c0e28ad11d7c86cf4ef9 not found: ID does not exist" containerID="522692386d728b8dce9d5bb112e19798cd1d98d5b857c0e28ad11d7c86cf4ef9" Dec 06 16:27:11 crc kubenswrapper[4813]: I1206 16:27:11.268747 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"522692386d728b8dce9d5bb112e19798cd1d98d5b857c0e28ad11d7c86cf4ef9"} err="failed to get container status \"522692386d728b8dce9d5bb112e19798cd1d98d5b857c0e28ad11d7c86cf4ef9\": rpc error: code = NotFound desc = could not find container \"522692386d728b8dce9d5bb112e19798cd1d98d5b857c0e28ad11d7c86cf4ef9\": container with ID starting with 522692386d728b8dce9d5bb112e19798cd1d98d5b857c0e28ad11d7c86cf4ef9 not found: ID does not exist" Dec 06 16:27:11 crc kubenswrapper[4813]: I1206 16:27:11.268774 4813 scope.go:117] "RemoveContainer" containerID="9df454d2b2605163125c03671cb312630b20cdac92eef3791d37b1317bb2af76" Dec 06 16:27:11 crc kubenswrapper[4813]: E1206 16:27:11.269360 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9df454d2b2605163125c03671cb312630b20cdac92eef3791d37b1317bb2af76\": container with ID starting with 9df454d2b2605163125c03671cb312630b20cdac92eef3791d37b1317bb2af76 not found: ID does not exist" containerID="9df454d2b2605163125c03671cb312630b20cdac92eef3791d37b1317bb2af76" Dec 06 16:27:11 crc kubenswrapper[4813]: I1206 16:27:11.269403 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9df454d2b2605163125c03671cb312630b20cdac92eef3791d37b1317bb2af76"} err="failed to get container status \"9df454d2b2605163125c03671cb312630b20cdac92eef3791d37b1317bb2af76\": rpc error: code = NotFound desc = could not find container \"9df454d2b2605163125c03671cb312630b20cdac92eef3791d37b1317bb2af76\": container with ID starting with 9df454d2b2605163125c03671cb312630b20cdac92eef3791d37b1317bb2af76 not found: ID does not exist" Dec 06 16:27:12 crc kubenswrapper[4813]: I1206 16:27:12.499250 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d049dac5-fc23-4b86-ab68-c114ed0b6371" path="/var/lib/kubelet/pods/d049dac5-fc23-4b86-ab68-c114ed0b6371/volumes" Dec 06 16:27:17 crc kubenswrapper[4813]: I1206 16:27:17.487590 4813 scope.go:117] "RemoveContainer" containerID="1b8cc98443ea491aa620632d54d48664f2e8cd65b0790f1013b24051a06625ca" Dec 06 16:27:17 crc kubenswrapper[4813]: E1206 16:27:17.488572 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:27:31 crc kubenswrapper[4813]: I1206 16:27:31.487376 4813 scope.go:117] "RemoveContainer" containerID="1b8cc98443ea491aa620632d54d48664f2e8cd65b0790f1013b24051a06625ca" Dec 06 16:27:31 crc kubenswrapper[4813]: E1206 16:27:31.488110 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:27:45 crc kubenswrapper[4813]: I1206 16:27:45.488925 4813 scope.go:117] "RemoveContainer" containerID="1b8cc98443ea491aa620632d54d48664f2e8cd65b0790f1013b24051a06625ca" Dec 06 16:27:45 crc kubenswrapper[4813]: E1206 16:27:45.494153 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:27:57 crc kubenswrapper[4813]: I1206 16:27:57.487154 4813 scope.go:117] "RemoveContainer" containerID="1b8cc98443ea491aa620632d54d48664f2e8cd65b0790f1013b24051a06625ca" Dec 06 16:27:57 crc kubenswrapper[4813]: E1206 16:27:57.488438 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:28:10 crc kubenswrapper[4813]: I1206 16:28:10.496029 4813 scope.go:117] "RemoveContainer" containerID="1b8cc98443ea491aa620632d54d48664f2e8cd65b0790f1013b24051a06625ca" Dec 06 16:28:10 crc kubenswrapper[4813]: E1206 16:28:10.497228 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:28:23 crc kubenswrapper[4813]: I1206 16:28:23.487251 4813 scope.go:117] "RemoveContainer" containerID="1b8cc98443ea491aa620632d54d48664f2e8cd65b0790f1013b24051a06625ca" Dec 06 16:28:23 crc kubenswrapper[4813]: E1206 16:28:23.488025 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:28:36 crc kubenswrapper[4813]: I1206 16:28:36.486871 4813 scope.go:117] "RemoveContainer" containerID="1b8cc98443ea491aa620632d54d48664f2e8cd65b0790f1013b24051a06625ca" Dec 06 16:28:36 crc kubenswrapper[4813]: E1206 16:28:36.488425 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:28:48 crc kubenswrapper[4813]: I1206 16:28:48.186853 4813 generic.go:334] "Generic (PLEG): container finished" podID="7ba82e65-8192-4943-8a2a-863d041ad231" containerID="a68abb34910726f382e7410aabbe8b2851f441e06609b7ea07f5e963ca5195e8" exitCode=0 Dec 06 16:28:48 crc kubenswrapper[4813]: I1206 16:28:48.186935 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" event={"ID":"7ba82e65-8192-4943-8a2a-863d041ad231","Type":"ContainerDied","Data":"a68abb34910726f382e7410aabbe8b2851f441e06609b7ea07f5e963ca5195e8"} Dec 06 16:28:49 crc kubenswrapper[4813]: I1206 16:28:49.701962 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:28:49 crc kubenswrapper[4813]: I1206 16:28:49.730116 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-nova-migration-ssh-key-0\") pod \"7ba82e65-8192-4943-8a2a-863d041ad231\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " Dec 06 16:28:49 crc kubenswrapper[4813]: I1206 16:28:49.732509 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-nova-cell1-compute-config-1\") pod \"7ba82e65-8192-4943-8a2a-863d041ad231\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " Dec 06 16:28:49 crc kubenswrapper[4813]: I1206 16:28:49.732750 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-nova-combined-ca-bundle\") pod \"7ba82e65-8192-4943-8a2a-863d041ad231\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " Dec 06 16:28:49 crc kubenswrapper[4813]: I1206 16:28:49.732903 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-inventory\") pod \"7ba82e65-8192-4943-8a2a-863d041ad231\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " Dec 06 16:28:49 crc kubenswrapper[4813]: I1206 16:28:49.732967 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-ssh-key\") pod \"7ba82e65-8192-4943-8a2a-863d041ad231\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " Dec 06 16:28:49 crc kubenswrapper[4813]: I1206 16:28:49.733033 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mz72t\" (UniqueName: \"kubernetes.io/projected/7ba82e65-8192-4943-8a2a-863d041ad231-kube-api-access-mz72t\") pod \"7ba82e65-8192-4943-8a2a-863d041ad231\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " Dec 06 16:28:49 crc kubenswrapper[4813]: I1206 16:28:49.733070 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-nova-cell1-compute-config-0\") pod \"7ba82e65-8192-4943-8a2a-863d041ad231\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " Dec 06 16:28:49 crc kubenswrapper[4813]: I1206 16:28:49.733094 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7ba82e65-8192-4943-8a2a-863d041ad231-nova-extra-config-0\") pod \"7ba82e65-8192-4943-8a2a-863d041ad231\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " Dec 06 16:28:49 crc kubenswrapper[4813]: I1206 16:28:49.733127 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-nova-migration-ssh-key-1\") pod \"7ba82e65-8192-4943-8a2a-863d041ad231\" (UID: \"7ba82e65-8192-4943-8a2a-863d041ad231\") " Dec 06 16:28:49 crc kubenswrapper[4813]: I1206 16:28:49.749689 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "7ba82e65-8192-4943-8a2a-863d041ad231" (UID: "7ba82e65-8192-4943-8a2a-863d041ad231"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:28:49 crc kubenswrapper[4813]: I1206 16:28:49.754831 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ba82e65-8192-4943-8a2a-863d041ad231-kube-api-access-mz72t" (OuterVolumeSpecName: "kube-api-access-mz72t") pod "7ba82e65-8192-4943-8a2a-863d041ad231" (UID: "7ba82e65-8192-4943-8a2a-863d041ad231"). InnerVolumeSpecName "kube-api-access-mz72t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:28:49 crc kubenswrapper[4813]: I1206 16:28:49.784937 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ba82e65-8192-4943-8a2a-863d041ad231-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "7ba82e65-8192-4943-8a2a-863d041ad231" (UID: "7ba82e65-8192-4943-8a2a-863d041ad231"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:28:49 crc kubenswrapper[4813]: I1206 16:28:49.808414 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "7ba82e65-8192-4943-8a2a-863d041ad231" (UID: "7ba82e65-8192-4943-8a2a-863d041ad231"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:28:49 crc kubenswrapper[4813]: I1206 16:28:49.808458 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "7ba82e65-8192-4943-8a2a-863d041ad231" (UID: "7ba82e65-8192-4943-8a2a-863d041ad231"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:28:49 crc kubenswrapper[4813]: I1206 16:28:49.813561 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "7ba82e65-8192-4943-8a2a-863d041ad231" (UID: "7ba82e65-8192-4943-8a2a-863d041ad231"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:28:49 crc kubenswrapper[4813]: I1206 16:28:49.821567 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-inventory" (OuterVolumeSpecName: "inventory") pod "7ba82e65-8192-4943-8a2a-863d041ad231" (UID: "7ba82e65-8192-4943-8a2a-863d041ad231"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:28:49 crc kubenswrapper[4813]: I1206 16:28:49.828253 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "7ba82e65-8192-4943-8a2a-863d041ad231" (UID: "7ba82e65-8192-4943-8a2a-863d041ad231"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:28:49 crc kubenswrapper[4813]: I1206 16:28:49.828781 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7ba82e65-8192-4943-8a2a-863d041ad231" (UID: "7ba82e65-8192-4943-8a2a-863d041ad231"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:28:49 crc kubenswrapper[4813]: I1206 16:28:49.836533 4813 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 06 16:28:49 crc kubenswrapper[4813]: I1206 16:28:49.836569 4813 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:28:49 crc kubenswrapper[4813]: I1206 16:28:49.836580 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 16:28:49 crc kubenswrapper[4813]: I1206 16:28:49.836589 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 16:28:49 crc kubenswrapper[4813]: I1206 16:28:49.836599 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mz72t\" (UniqueName: \"kubernetes.io/projected/7ba82e65-8192-4943-8a2a-863d041ad231-kube-api-access-mz72t\") on node \"crc\" DevicePath \"\"" Dec 06 16:28:49 crc kubenswrapper[4813]: I1206 16:28:49.836647 4813 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7ba82e65-8192-4943-8a2a-863d041ad231-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 16:28:49 crc kubenswrapper[4813]: I1206 16:28:49.836657 4813 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 16:28:49 crc kubenswrapper[4813]: I1206 16:28:49.836665 4813 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 06 16:28:49 crc kubenswrapper[4813]: I1206 16:28:49.836674 4813 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7ba82e65-8192-4943-8a2a-863d041ad231-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.216072 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" event={"ID":"7ba82e65-8192-4943-8a2a-863d041ad231","Type":"ContainerDied","Data":"4550799567e52d98805e1acc6be838dcbf81b8c05a15852f901b87bd9e20da23"} Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.216106 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4550799567e52d98805e1acc6be838dcbf81b8c05a15852f901b87bd9e20da23" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.216170 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mfh5s" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.339661 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt"] Dec 06 16:28:50 crc kubenswrapper[4813]: E1206 16:28:50.340025 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d049dac5-fc23-4b86-ab68-c114ed0b6371" containerName="registry-server" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.340043 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d049dac5-fc23-4b86-ab68-c114ed0b6371" containerName="registry-server" Dec 06 16:28:50 crc kubenswrapper[4813]: E1206 16:28:50.340065 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d049dac5-fc23-4b86-ab68-c114ed0b6371" containerName="extract-utilities" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.340072 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d049dac5-fc23-4b86-ab68-c114ed0b6371" containerName="extract-utilities" Dec 06 16:28:50 crc kubenswrapper[4813]: E1206 16:28:50.340078 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d049dac5-fc23-4b86-ab68-c114ed0b6371" containerName="extract-content" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.340084 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d049dac5-fc23-4b86-ab68-c114ed0b6371" containerName="extract-content" Dec 06 16:28:50 crc kubenswrapper[4813]: E1206 16:28:50.340098 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ba82e65-8192-4943-8a2a-863d041ad231" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.340104 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ba82e65-8192-4943-8a2a-863d041ad231" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.340299 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d049dac5-fc23-4b86-ab68-c114ed0b6371" containerName="registry-server" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.340314 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ba82e65-8192-4943-8a2a-863d041ad231" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.340995 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.346493 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.348933 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.349008 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.349064 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.350418 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jvkhf" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.364466 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt"] Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.456064 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt\" (UID: \"a1ecf1f7-01c3-456c-8698-1353e0651c24\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.456345 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt\" (UID: \"a1ecf1f7-01c3-456c-8698-1353e0651c24\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.456425 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt\" (UID: \"a1ecf1f7-01c3-456c-8698-1353e0651c24\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.456455 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt\" (UID: \"a1ecf1f7-01c3-456c-8698-1353e0651c24\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.456473 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mv77h\" (UniqueName: \"kubernetes.io/projected/a1ecf1f7-01c3-456c-8698-1353e0651c24-kube-api-access-mv77h\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt\" (UID: \"a1ecf1f7-01c3-456c-8698-1353e0651c24\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.456493 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt\" (UID: \"a1ecf1f7-01c3-456c-8698-1353e0651c24\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.456511 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt\" (UID: \"a1ecf1f7-01c3-456c-8698-1353e0651c24\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.492612 4813 scope.go:117] "RemoveContainer" containerID="1b8cc98443ea491aa620632d54d48664f2e8cd65b0790f1013b24051a06625ca" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.558929 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt\" (UID: \"a1ecf1f7-01c3-456c-8698-1353e0651c24\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.559105 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt\" (UID: \"a1ecf1f7-01c3-456c-8698-1353e0651c24\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.559182 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mv77h\" (UniqueName: \"kubernetes.io/projected/a1ecf1f7-01c3-456c-8698-1353e0651c24-kube-api-access-mv77h\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt\" (UID: \"a1ecf1f7-01c3-456c-8698-1353e0651c24\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.559206 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt\" (UID: \"a1ecf1f7-01c3-456c-8698-1353e0651c24\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.559227 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt\" (UID: \"a1ecf1f7-01c3-456c-8698-1353e0651c24\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.559340 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt\" (UID: \"a1ecf1f7-01c3-456c-8698-1353e0651c24\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.559421 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt\" (UID: \"a1ecf1f7-01c3-456c-8698-1353e0651c24\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.562636 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt\" (UID: \"a1ecf1f7-01c3-456c-8698-1353e0651c24\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.562822 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt\" (UID: \"a1ecf1f7-01c3-456c-8698-1353e0651c24\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.563667 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt\" (UID: \"a1ecf1f7-01c3-456c-8698-1353e0651c24\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.564521 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt\" (UID: \"a1ecf1f7-01c3-456c-8698-1353e0651c24\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.565313 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt\" (UID: \"a1ecf1f7-01c3-456c-8698-1353e0651c24\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.569345 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt\" (UID: \"a1ecf1f7-01c3-456c-8698-1353e0651c24\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.590934 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mv77h\" (UniqueName: \"kubernetes.io/projected/a1ecf1f7-01c3-456c-8698-1353e0651c24-kube-api-access-mv77h\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt\" (UID: \"a1ecf1f7-01c3-456c-8698-1353e0651c24\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt" Dec 06 16:28:50 crc kubenswrapper[4813]: I1206 16:28:50.659088 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt" Dec 06 16:28:51 crc kubenswrapper[4813]: I1206 16:28:51.199005 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt"] Dec 06 16:28:51 crc kubenswrapper[4813]: W1206 16:28:51.209443 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1ecf1f7_01c3_456c_8698_1353e0651c24.slice/crio-1f7c6387ab6c97d5760c8668ef9db0688bfd20980c02f06bc1a1315bba6bae1b WatchSource:0}: Error finding container 1f7c6387ab6c97d5760c8668ef9db0688bfd20980c02f06bc1a1315bba6bae1b: Status 404 returned error can't find the container with id 1f7c6387ab6c97d5760c8668ef9db0688bfd20980c02f06bc1a1315bba6bae1b Dec 06 16:28:51 crc kubenswrapper[4813]: I1206 16:28:51.212205 4813 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 16:28:51 crc kubenswrapper[4813]: I1206 16:28:51.229285 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerStarted","Data":"0f04b80ef02ad2942e19db0f6e8b08273d91a189f92b54348561c653b70e0191"} Dec 06 16:28:51 crc kubenswrapper[4813]: I1206 16:28:51.230366 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt" event={"ID":"a1ecf1f7-01c3-456c-8698-1353e0651c24","Type":"ContainerStarted","Data":"1f7c6387ab6c97d5760c8668ef9db0688bfd20980c02f06bc1a1315bba6bae1b"} Dec 06 16:28:52 crc kubenswrapper[4813]: I1206 16:28:52.239716 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt" event={"ID":"a1ecf1f7-01c3-456c-8698-1353e0651c24","Type":"ContainerStarted","Data":"80d1bd4559c837d91d588b5fd7552e39f07c3843233eaad19ed218a86844f6c9"} Dec 06 16:28:52 crc kubenswrapper[4813]: I1206 16:28:52.260640 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt" podStartSLOduration=1.786830949 podStartE2EDuration="2.260624121s" podCreationTimestamp="2025-12-06 16:28:50 +0000 UTC" firstStartedPulling="2025-12-06 16:28:51.212015065 +0000 UTC m=+2571.102894641" lastFinishedPulling="2025-12-06 16:28:51.685808227 +0000 UTC m=+2571.576687813" observedRunningTime="2025-12-06 16:28:52.255058413 +0000 UTC m=+2572.145937989" watchObservedRunningTime="2025-12-06 16:28:52.260624121 +0000 UTC m=+2572.151503697" Dec 06 16:30:00 crc kubenswrapper[4813]: I1206 16:30:00.165586 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417310-d8z7c"] Dec 06 16:30:00 crc kubenswrapper[4813]: I1206 16:30:00.167804 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417310-d8z7c" Dec 06 16:30:00 crc kubenswrapper[4813]: I1206 16:30:00.174889 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 16:30:00 crc kubenswrapper[4813]: I1206 16:30:00.179489 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417310-d8z7c"] Dec 06 16:30:00 crc kubenswrapper[4813]: I1206 16:30:00.180563 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 16:30:00 crc kubenswrapper[4813]: I1206 16:30:00.272355 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl8lh\" (UniqueName: \"kubernetes.io/projected/c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3-kube-api-access-tl8lh\") pod \"collect-profiles-29417310-d8z7c\" (UID: \"c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417310-d8z7c" Dec 06 16:30:00 crc kubenswrapper[4813]: I1206 16:30:00.272399 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3-config-volume\") pod \"collect-profiles-29417310-d8z7c\" (UID: \"c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417310-d8z7c" Dec 06 16:30:00 crc kubenswrapper[4813]: I1206 16:30:00.272581 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3-secret-volume\") pod \"collect-profiles-29417310-d8z7c\" (UID: \"c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417310-d8z7c" Dec 06 16:30:00 crc kubenswrapper[4813]: I1206 16:30:00.374215 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl8lh\" (UniqueName: \"kubernetes.io/projected/c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3-kube-api-access-tl8lh\") pod \"collect-profiles-29417310-d8z7c\" (UID: \"c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417310-d8z7c" Dec 06 16:30:00 crc kubenswrapper[4813]: I1206 16:30:00.374256 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3-config-volume\") pod \"collect-profiles-29417310-d8z7c\" (UID: \"c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417310-d8z7c" Dec 06 16:30:00 crc kubenswrapper[4813]: I1206 16:30:00.374379 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3-secret-volume\") pod \"collect-profiles-29417310-d8z7c\" (UID: \"c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417310-d8z7c" Dec 06 16:30:00 crc kubenswrapper[4813]: I1206 16:30:00.375443 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3-config-volume\") pod \"collect-profiles-29417310-d8z7c\" (UID: \"c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417310-d8z7c" Dec 06 16:30:00 crc kubenswrapper[4813]: I1206 16:30:00.388476 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3-secret-volume\") pod \"collect-profiles-29417310-d8z7c\" (UID: \"c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417310-d8z7c" Dec 06 16:30:00 crc kubenswrapper[4813]: I1206 16:30:00.392840 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl8lh\" (UniqueName: \"kubernetes.io/projected/c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3-kube-api-access-tl8lh\") pod \"collect-profiles-29417310-d8z7c\" (UID: \"c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417310-d8z7c" Dec 06 16:30:00 crc kubenswrapper[4813]: I1206 16:30:00.491046 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 16:30:00 crc kubenswrapper[4813]: I1206 16:30:00.500474 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417310-d8z7c" Dec 06 16:30:01 crc kubenswrapper[4813]: I1206 16:30:01.006741 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417310-d8z7c"] Dec 06 16:30:02 crc kubenswrapper[4813]: I1206 16:30:02.012631 4813 generic.go:334] "Generic (PLEG): container finished" podID="c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3" containerID="9b5fd2093f5dcba72d01238778e20caf76472dd86e71a49a4064de2345717704" exitCode=0 Dec 06 16:30:02 crc kubenswrapper[4813]: I1206 16:30:02.012817 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417310-d8z7c" event={"ID":"c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3","Type":"ContainerDied","Data":"9b5fd2093f5dcba72d01238778e20caf76472dd86e71a49a4064de2345717704"} Dec 06 16:30:02 crc kubenswrapper[4813]: I1206 16:30:02.013841 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417310-d8z7c" event={"ID":"c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3","Type":"ContainerStarted","Data":"736293ee2dfb08c60b243ecf9cc76ba53798072ac6bb31f4222cf8710566c0ed"} Dec 06 16:30:03 crc kubenswrapper[4813]: I1206 16:30:03.478294 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417310-d8z7c" Dec 06 16:30:03 crc kubenswrapper[4813]: I1206 16:30:03.561766 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3-secret-volume\") pod \"c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3\" (UID: \"c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3\") " Dec 06 16:30:03 crc kubenswrapper[4813]: I1206 16:30:03.561877 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3-config-volume\") pod \"c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3\" (UID: \"c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3\") " Dec 06 16:30:03 crc kubenswrapper[4813]: I1206 16:30:03.562073 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tl8lh\" (UniqueName: \"kubernetes.io/projected/c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3-kube-api-access-tl8lh\") pod \"c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3\" (UID: \"c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3\") " Dec 06 16:30:03 crc kubenswrapper[4813]: I1206 16:30:03.562850 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3-config-volume" (OuterVolumeSpecName: "config-volume") pod "c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3" (UID: "c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:30:03 crc kubenswrapper[4813]: I1206 16:30:03.568232 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3-kube-api-access-tl8lh" (OuterVolumeSpecName: "kube-api-access-tl8lh") pod "c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3" (UID: "c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3"). InnerVolumeSpecName "kube-api-access-tl8lh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:30:03 crc kubenswrapper[4813]: I1206 16:30:03.572996 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3" (UID: "c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:30:03 crc kubenswrapper[4813]: I1206 16:30:03.665024 4813 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 16:30:03 crc kubenswrapper[4813]: I1206 16:30:03.665065 4813 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 16:30:03 crc kubenswrapper[4813]: I1206 16:30:03.665079 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tl8lh\" (UniqueName: \"kubernetes.io/projected/c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3-kube-api-access-tl8lh\") on node \"crc\" DevicePath \"\"" Dec 06 16:30:04 crc kubenswrapper[4813]: I1206 16:30:04.036373 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417310-d8z7c" event={"ID":"c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3","Type":"ContainerDied","Data":"736293ee2dfb08c60b243ecf9cc76ba53798072ac6bb31f4222cf8710566c0ed"} Dec 06 16:30:04 crc kubenswrapper[4813]: I1206 16:30:04.036430 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="736293ee2dfb08c60b243ecf9cc76ba53798072ac6bb31f4222cf8710566c0ed" Dec 06 16:30:04 crc kubenswrapper[4813]: I1206 16:30:04.036722 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417310-d8z7c" Dec 06 16:30:04 crc kubenswrapper[4813]: I1206 16:30:04.605311 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417265-hcswl"] Dec 06 16:30:04 crc kubenswrapper[4813]: I1206 16:30:04.614494 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417265-hcswl"] Dec 06 16:30:06 crc kubenswrapper[4813]: I1206 16:30:06.496788 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8810f12c-6dbc-4bf9-b27e-29ebc5986955" path="/var/lib/kubelet/pods/8810f12c-6dbc-4bf9-b27e-29ebc5986955/volumes" Dec 06 16:30:25 crc kubenswrapper[4813]: I1206 16:30:25.059704 4813 scope.go:117] "RemoveContainer" containerID="2fba0692193bb5fa545d890a31ff21e880014c4c26c01164df7322b66f8ec130" Dec 06 16:31:19 crc kubenswrapper[4813]: I1206 16:31:19.427602 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:31:19 crc kubenswrapper[4813]: I1206 16:31:19.428232 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:31:49 crc kubenswrapper[4813]: I1206 16:31:49.428046 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:31:49 crc kubenswrapper[4813]: I1206 16:31:49.429831 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:31:49 crc kubenswrapper[4813]: I1206 16:31:49.469077 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7452r"] Dec 06 16:31:49 crc kubenswrapper[4813]: E1206 16:31:49.469597 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3" containerName="collect-profiles" Dec 06 16:31:49 crc kubenswrapper[4813]: I1206 16:31:49.469619 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3" containerName="collect-profiles" Dec 06 16:31:49 crc kubenswrapper[4813]: I1206 16:31:49.469916 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2aa83d5-d83e-4a2d-9a2c-74d2fec8cde3" containerName="collect-profiles" Dec 06 16:31:49 crc kubenswrapper[4813]: I1206 16:31:49.471577 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7452r" Dec 06 16:31:49 crc kubenswrapper[4813]: I1206 16:31:49.489336 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7452r"] Dec 06 16:31:49 crc kubenswrapper[4813]: I1206 16:31:49.563390 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c426e12-23d7-4821-bdc3-882990a9f41e-catalog-content\") pod \"community-operators-7452r\" (UID: \"5c426e12-23d7-4821-bdc3-882990a9f41e\") " pod="openshift-marketplace/community-operators-7452r" Dec 06 16:31:49 crc kubenswrapper[4813]: I1206 16:31:49.563695 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c426e12-23d7-4821-bdc3-882990a9f41e-utilities\") pod \"community-operators-7452r\" (UID: \"5c426e12-23d7-4821-bdc3-882990a9f41e\") " pod="openshift-marketplace/community-operators-7452r" Dec 06 16:31:49 crc kubenswrapper[4813]: I1206 16:31:49.563890 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62ssb\" (UniqueName: \"kubernetes.io/projected/5c426e12-23d7-4821-bdc3-882990a9f41e-kube-api-access-62ssb\") pod \"community-operators-7452r\" (UID: \"5c426e12-23d7-4821-bdc3-882990a9f41e\") " pod="openshift-marketplace/community-operators-7452r" Dec 06 16:31:49 crc kubenswrapper[4813]: I1206 16:31:49.666325 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c426e12-23d7-4821-bdc3-882990a9f41e-catalog-content\") pod \"community-operators-7452r\" (UID: \"5c426e12-23d7-4821-bdc3-882990a9f41e\") " pod="openshift-marketplace/community-operators-7452r" Dec 06 16:31:49 crc kubenswrapper[4813]: I1206 16:31:49.666421 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c426e12-23d7-4821-bdc3-882990a9f41e-utilities\") pod \"community-operators-7452r\" (UID: \"5c426e12-23d7-4821-bdc3-882990a9f41e\") " pod="openshift-marketplace/community-operators-7452r" Dec 06 16:31:49 crc kubenswrapper[4813]: I1206 16:31:49.666539 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62ssb\" (UniqueName: \"kubernetes.io/projected/5c426e12-23d7-4821-bdc3-882990a9f41e-kube-api-access-62ssb\") pod \"community-operators-7452r\" (UID: \"5c426e12-23d7-4821-bdc3-882990a9f41e\") " pod="openshift-marketplace/community-operators-7452r" Dec 06 16:31:49 crc kubenswrapper[4813]: I1206 16:31:49.666986 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c426e12-23d7-4821-bdc3-882990a9f41e-catalog-content\") pod \"community-operators-7452r\" (UID: \"5c426e12-23d7-4821-bdc3-882990a9f41e\") " pod="openshift-marketplace/community-operators-7452r" Dec 06 16:31:49 crc kubenswrapper[4813]: I1206 16:31:49.667189 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c426e12-23d7-4821-bdc3-882990a9f41e-utilities\") pod \"community-operators-7452r\" (UID: \"5c426e12-23d7-4821-bdc3-882990a9f41e\") " pod="openshift-marketplace/community-operators-7452r" Dec 06 16:31:49 crc kubenswrapper[4813]: I1206 16:31:49.686074 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62ssb\" (UniqueName: \"kubernetes.io/projected/5c426e12-23d7-4821-bdc3-882990a9f41e-kube-api-access-62ssb\") pod \"community-operators-7452r\" (UID: \"5c426e12-23d7-4821-bdc3-882990a9f41e\") " pod="openshift-marketplace/community-operators-7452r" Dec 06 16:31:49 crc kubenswrapper[4813]: I1206 16:31:49.806573 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7452r" Dec 06 16:31:50 crc kubenswrapper[4813]: I1206 16:31:50.420548 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7452r"] Dec 06 16:31:51 crc kubenswrapper[4813]: I1206 16:31:51.170473 4813 generic.go:334] "Generic (PLEG): container finished" podID="5c426e12-23d7-4821-bdc3-882990a9f41e" containerID="75c955ce0d5a8456ca005ae2f363da4277e17d813b8bb007e254154828141ba9" exitCode=0 Dec 06 16:31:51 crc kubenswrapper[4813]: I1206 16:31:51.170962 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7452r" event={"ID":"5c426e12-23d7-4821-bdc3-882990a9f41e","Type":"ContainerDied","Data":"75c955ce0d5a8456ca005ae2f363da4277e17d813b8bb007e254154828141ba9"} Dec 06 16:31:51 crc kubenswrapper[4813]: I1206 16:31:51.171032 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7452r" event={"ID":"5c426e12-23d7-4821-bdc3-882990a9f41e","Type":"ContainerStarted","Data":"5913921de8c35f0d18ebce5103adefe45d8f223b00e22b023ee485f02f2f1708"} Dec 06 16:31:52 crc kubenswrapper[4813]: I1206 16:31:52.179512 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7452r" event={"ID":"5c426e12-23d7-4821-bdc3-882990a9f41e","Type":"ContainerStarted","Data":"30e5fc5cf460b1f91911f313f7c6ccc0935930b8cf79615f6393160c78b7f10a"} Dec 06 16:31:54 crc kubenswrapper[4813]: I1206 16:31:54.206986 4813 generic.go:334] "Generic (PLEG): container finished" podID="5c426e12-23d7-4821-bdc3-882990a9f41e" containerID="30e5fc5cf460b1f91911f313f7c6ccc0935930b8cf79615f6393160c78b7f10a" exitCode=0 Dec 06 16:31:54 crc kubenswrapper[4813]: I1206 16:31:54.207401 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7452r" event={"ID":"5c426e12-23d7-4821-bdc3-882990a9f41e","Type":"ContainerDied","Data":"30e5fc5cf460b1f91911f313f7c6ccc0935930b8cf79615f6393160c78b7f10a"} Dec 06 16:31:55 crc kubenswrapper[4813]: I1206 16:31:55.223079 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7452r" event={"ID":"5c426e12-23d7-4821-bdc3-882990a9f41e","Type":"ContainerStarted","Data":"497055b9b13dcdaea2551b9a0cc1cc221e0750445e1d6df16ba7a5ccc060f104"} Dec 06 16:31:55 crc kubenswrapper[4813]: I1206 16:31:55.245660 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-q4tb6"] Dec 06 16:31:55 crc kubenswrapper[4813]: I1206 16:31:55.251280 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q4tb6" Dec 06 16:31:55 crc kubenswrapper[4813]: I1206 16:31:55.263250 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q4tb6"] Dec 06 16:31:55 crc kubenswrapper[4813]: I1206 16:31:55.264858 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7452r" podStartSLOduration=2.816444041 podStartE2EDuration="6.264840611s" podCreationTimestamp="2025-12-06 16:31:49 +0000 UTC" firstStartedPulling="2025-12-06 16:31:51.172781664 +0000 UTC m=+2751.063661240" lastFinishedPulling="2025-12-06 16:31:54.621178194 +0000 UTC m=+2754.512057810" observedRunningTime="2025-12-06 16:31:55.261595256 +0000 UTC m=+2755.152474832" watchObservedRunningTime="2025-12-06 16:31:55.264840611 +0000 UTC m=+2755.155720187" Dec 06 16:31:55 crc kubenswrapper[4813]: I1206 16:31:55.325140 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksgdw\" (UniqueName: \"kubernetes.io/projected/7ee2a239-1568-4b54-9196-af436b40399c-kube-api-access-ksgdw\") pod \"redhat-marketplace-q4tb6\" (UID: \"7ee2a239-1568-4b54-9196-af436b40399c\") " pod="openshift-marketplace/redhat-marketplace-q4tb6" Dec 06 16:31:55 crc kubenswrapper[4813]: I1206 16:31:55.325360 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ee2a239-1568-4b54-9196-af436b40399c-catalog-content\") pod \"redhat-marketplace-q4tb6\" (UID: \"7ee2a239-1568-4b54-9196-af436b40399c\") " pod="openshift-marketplace/redhat-marketplace-q4tb6" Dec 06 16:31:55 crc kubenswrapper[4813]: I1206 16:31:55.325624 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ee2a239-1568-4b54-9196-af436b40399c-utilities\") pod \"redhat-marketplace-q4tb6\" (UID: \"7ee2a239-1568-4b54-9196-af436b40399c\") " pod="openshift-marketplace/redhat-marketplace-q4tb6" Dec 06 16:31:55 crc kubenswrapper[4813]: I1206 16:31:55.428025 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksgdw\" (UniqueName: \"kubernetes.io/projected/7ee2a239-1568-4b54-9196-af436b40399c-kube-api-access-ksgdw\") pod \"redhat-marketplace-q4tb6\" (UID: \"7ee2a239-1568-4b54-9196-af436b40399c\") " pod="openshift-marketplace/redhat-marketplace-q4tb6" Dec 06 16:31:55 crc kubenswrapper[4813]: I1206 16:31:55.428206 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ee2a239-1568-4b54-9196-af436b40399c-catalog-content\") pod \"redhat-marketplace-q4tb6\" (UID: \"7ee2a239-1568-4b54-9196-af436b40399c\") " pod="openshift-marketplace/redhat-marketplace-q4tb6" Dec 06 16:31:55 crc kubenswrapper[4813]: I1206 16:31:55.428397 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ee2a239-1568-4b54-9196-af436b40399c-utilities\") pod \"redhat-marketplace-q4tb6\" (UID: \"7ee2a239-1568-4b54-9196-af436b40399c\") " pod="openshift-marketplace/redhat-marketplace-q4tb6" Dec 06 16:31:55 crc kubenswrapper[4813]: I1206 16:31:55.428778 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ee2a239-1568-4b54-9196-af436b40399c-catalog-content\") pod \"redhat-marketplace-q4tb6\" (UID: \"7ee2a239-1568-4b54-9196-af436b40399c\") " pod="openshift-marketplace/redhat-marketplace-q4tb6" Dec 06 16:31:55 crc kubenswrapper[4813]: I1206 16:31:55.428827 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ee2a239-1568-4b54-9196-af436b40399c-utilities\") pod \"redhat-marketplace-q4tb6\" (UID: \"7ee2a239-1568-4b54-9196-af436b40399c\") " pod="openshift-marketplace/redhat-marketplace-q4tb6" Dec 06 16:31:55 crc kubenswrapper[4813]: I1206 16:31:55.450488 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksgdw\" (UniqueName: \"kubernetes.io/projected/7ee2a239-1568-4b54-9196-af436b40399c-kube-api-access-ksgdw\") pod \"redhat-marketplace-q4tb6\" (UID: \"7ee2a239-1568-4b54-9196-af436b40399c\") " pod="openshift-marketplace/redhat-marketplace-q4tb6" Dec 06 16:31:55 crc kubenswrapper[4813]: I1206 16:31:55.567310 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q4tb6" Dec 06 16:31:56 crc kubenswrapper[4813]: I1206 16:31:56.065033 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q4tb6"] Dec 06 16:31:56 crc kubenswrapper[4813]: I1206 16:31:56.239956 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4tb6" event={"ID":"7ee2a239-1568-4b54-9196-af436b40399c","Type":"ContainerStarted","Data":"b2f483a7b048cf16b42e9b6b4eb90f0120f12fe92fe0e56e477cef2193d8edf5"} Dec 06 16:31:57 crc kubenswrapper[4813]: I1206 16:31:57.250870 4813 generic.go:334] "Generic (PLEG): container finished" podID="7ee2a239-1568-4b54-9196-af436b40399c" containerID="187b5398f8c5398e130ffeb4f4d7b132a3721e70243bad31b0e0d8b66adae2eb" exitCode=0 Dec 06 16:31:57 crc kubenswrapper[4813]: I1206 16:31:57.251109 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4tb6" event={"ID":"7ee2a239-1568-4b54-9196-af436b40399c","Type":"ContainerDied","Data":"187b5398f8c5398e130ffeb4f4d7b132a3721e70243bad31b0e0d8b66adae2eb"} Dec 06 16:31:58 crc kubenswrapper[4813]: I1206 16:31:58.262602 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4tb6" event={"ID":"7ee2a239-1568-4b54-9196-af436b40399c","Type":"ContainerStarted","Data":"1b9b388415e6f3aae84e93974a436e888daa72487976c183f81111ed89f56863"} Dec 06 16:31:59 crc kubenswrapper[4813]: I1206 16:31:59.282972 4813 generic.go:334] "Generic (PLEG): container finished" podID="7ee2a239-1568-4b54-9196-af436b40399c" containerID="1b9b388415e6f3aae84e93974a436e888daa72487976c183f81111ed89f56863" exitCode=0 Dec 06 16:31:59 crc kubenswrapper[4813]: I1206 16:31:59.283573 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4tb6" event={"ID":"7ee2a239-1568-4b54-9196-af436b40399c","Type":"ContainerDied","Data":"1b9b388415e6f3aae84e93974a436e888daa72487976c183f81111ed89f56863"} Dec 06 16:31:59 crc kubenswrapper[4813]: I1206 16:31:59.807147 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7452r" Dec 06 16:31:59 crc kubenswrapper[4813]: I1206 16:31:59.807579 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7452r" Dec 06 16:31:59 crc kubenswrapper[4813]: I1206 16:31:59.880429 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7452r" Dec 06 16:32:00 crc kubenswrapper[4813]: I1206 16:32:00.298769 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4tb6" event={"ID":"7ee2a239-1568-4b54-9196-af436b40399c","Type":"ContainerStarted","Data":"7f9c405e09265913f0f7f8457dec061406d0e6f55e78e0ddd608f016482ee218"} Dec 06 16:32:00 crc kubenswrapper[4813]: I1206 16:32:00.329213 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-q4tb6" podStartSLOduration=2.628707404 podStartE2EDuration="5.329190187s" podCreationTimestamp="2025-12-06 16:31:55 +0000 UTC" firstStartedPulling="2025-12-06 16:31:57.253293418 +0000 UTC m=+2757.144173004" lastFinishedPulling="2025-12-06 16:31:59.953776171 +0000 UTC m=+2759.844655787" observedRunningTime="2025-12-06 16:32:00.318710403 +0000 UTC m=+2760.209589979" watchObservedRunningTime="2025-12-06 16:32:00.329190187 +0000 UTC m=+2760.220069773" Dec 06 16:32:00 crc kubenswrapper[4813]: I1206 16:32:00.356468 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7452r" Dec 06 16:32:01 crc kubenswrapper[4813]: I1206 16:32:01.310768 4813 generic.go:334] "Generic (PLEG): container finished" podID="a1ecf1f7-01c3-456c-8698-1353e0651c24" containerID="80d1bd4559c837d91d588b5fd7552e39f07c3843233eaad19ed218a86844f6c9" exitCode=0 Dec 06 16:32:01 crc kubenswrapper[4813]: I1206 16:32:01.312152 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt" event={"ID":"a1ecf1f7-01c3-456c-8698-1353e0651c24","Type":"ContainerDied","Data":"80d1bd4559c837d91d588b5fd7552e39f07c3843233eaad19ed218a86844f6c9"} Dec 06 16:32:02 crc kubenswrapper[4813]: I1206 16:32:02.229333 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7452r"] Dec 06 16:32:02 crc kubenswrapper[4813]: I1206 16:32:02.780876 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt" Dec 06 16:32:02 crc kubenswrapper[4813]: I1206 16:32:02.899500 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-ceilometer-compute-config-data-0\") pod \"a1ecf1f7-01c3-456c-8698-1353e0651c24\" (UID: \"a1ecf1f7-01c3-456c-8698-1353e0651c24\") " Dec 06 16:32:02 crc kubenswrapper[4813]: I1206 16:32:02.899745 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mv77h\" (UniqueName: \"kubernetes.io/projected/a1ecf1f7-01c3-456c-8698-1353e0651c24-kube-api-access-mv77h\") pod \"a1ecf1f7-01c3-456c-8698-1353e0651c24\" (UID: \"a1ecf1f7-01c3-456c-8698-1353e0651c24\") " Dec 06 16:32:02 crc kubenswrapper[4813]: I1206 16:32:02.899802 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-telemetry-combined-ca-bundle\") pod \"a1ecf1f7-01c3-456c-8698-1353e0651c24\" (UID: \"a1ecf1f7-01c3-456c-8698-1353e0651c24\") " Dec 06 16:32:02 crc kubenswrapper[4813]: I1206 16:32:02.899848 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-inventory\") pod \"a1ecf1f7-01c3-456c-8698-1353e0651c24\" (UID: \"a1ecf1f7-01c3-456c-8698-1353e0651c24\") " Dec 06 16:32:02 crc kubenswrapper[4813]: I1206 16:32:02.899913 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-ceilometer-compute-config-data-1\") pod \"a1ecf1f7-01c3-456c-8698-1353e0651c24\" (UID: \"a1ecf1f7-01c3-456c-8698-1353e0651c24\") " Dec 06 16:32:02 crc kubenswrapper[4813]: I1206 16:32:02.899938 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-ceilometer-compute-config-data-2\") pod \"a1ecf1f7-01c3-456c-8698-1353e0651c24\" (UID: \"a1ecf1f7-01c3-456c-8698-1353e0651c24\") " Dec 06 16:32:02 crc kubenswrapper[4813]: I1206 16:32:02.899994 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-ssh-key\") pod \"a1ecf1f7-01c3-456c-8698-1353e0651c24\" (UID: \"a1ecf1f7-01c3-456c-8698-1353e0651c24\") " Dec 06 16:32:02 crc kubenswrapper[4813]: I1206 16:32:02.905711 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "a1ecf1f7-01c3-456c-8698-1353e0651c24" (UID: "a1ecf1f7-01c3-456c-8698-1353e0651c24"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:32:02 crc kubenswrapper[4813]: I1206 16:32:02.907560 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1ecf1f7-01c3-456c-8698-1353e0651c24-kube-api-access-mv77h" (OuterVolumeSpecName: "kube-api-access-mv77h") pod "a1ecf1f7-01c3-456c-8698-1353e0651c24" (UID: "a1ecf1f7-01c3-456c-8698-1353e0651c24"). InnerVolumeSpecName "kube-api-access-mv77h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:32:02 crc kubenswrapper[4813]: I1206 16:32:02.936468 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-inventory" (OuterVolumeSpecName: "inventory") pod "a1ecf1f7-01c3-456c-8698-1353e0651c24" (UID: "a1ecf1f7-01c3-456c-8698-1353e0651c24"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:32:02 crc kubenswrapper[4813]: I1206 16:32:02.939243 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "a1ecf1f7-01c3-456c-8698-1353e0651c24" (UID: "a1ecf1f7-01c3-456c-8698-1353e0651c24"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:32:02 crc kubenswrapper[4813]: I1206 16:32:02.939890 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "a1ecf1f7-01c3-456c-8698-1353e0651c24" (UID: "a1ecf1f7-01c3-456c-8698-1353e0651c24"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:32:02 crc kubenswrapper[4813]: I1206 16:32:02.953797 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "a1ecf1f7-01c3-456c-8698-1353e0651c24" (UID: "a1ecf1f7-01c3-456c-8698-1353e0651c24"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:32:02 crc kubenswrapper[4813]: I1206 16:32:02.964733 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a1ecf1f7-01c3-456c-8698-1353e0651c24" (UID: "a1ecf1f7-01c3-456c-8698-1353e0651c24"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:32:03 crc kubenswrapper[4813]: I1206 16:32:03.004246 4813 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 06 16:32:03 crc kubenswrapper[4813]: I1206 16:32:03.004459 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mv77h\" (UniqueName: \"kubernetes.io/projected/a1ecf1f7-01c3-456c-8698-1353e0651c24-kube-api-access-mv77h\") on node \"crc\" DevicePath \"\"" Dec 06 16:32:03 crc kubenswrapper[4813]: I1206 16:32:03.004600 4813 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 16:32:03 crc kubenswrapper[4813]: I1206 16:32:03.004728 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 16:32:03 crc kubenswrapper[4813]: I1206 16:32:03.005024 4813 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 06 16:32:03 crc kubenswrapper[4813]: I1206 16:32:03.005163 4813 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 06 16:32:03 crc kubenswrapper[4813]: I1206 16:32:03.005365 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a1ecf1f7-01c3-456c-8698-1353e0651c24-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 16:32:03 crc kubenswrapper[4813]: I1206 16:32:03.336579 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt" event={"ID":"a1ecf1f7-01c3-456c-8698-1353e0651c24","Type":"ContainerDied","Data":"1f7c6387ab6c97d5760c8668ef9db0688bfd20980c02f06bc1a1315bba6bae1b"} Dec 06 16:32:03 crc kubenswrapper[4813]: I1206 16:32:03.336663 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f7c6387ab6c97d5760c8668ef9db0688bfd20980c02f06bc1a1315bba6bae1b" Dec 06 16:32:03 crc kubenswrapper[4813]: I1206 16:32:03.336705 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7452r" podUID="5c426e12-23d7-4821-bdc3-882990a9f41e" containerName="registry-server" containerID="cri-o://497055b9b13dcdaea2551b9a0cc1cc221e0750445e1d6df16ba7a5ccc060f104" gracePeriod=2 Dec 06 16:32:03 crc kubenswrapper[4813]: I1206 16:32:03.337141 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt" Dec 06 16:32:03 crc kubenswrapper[4813]: I1206 16:32:03.861059 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7452r" Dec 06 16:32:04 crc kubenswrapper[4813]: I1206 16:32:04.025123 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c426e12-23d7-4821-bdc3-882990a9f41e-catalog-content\") pod \"5c426e12-23d7-4821-bdc3-882990a9f41e\" (UID: \"5c426e12-23d7-4821-bdc3-882990a9f41e\") " Dec 06 16:32:04 crc kubenswrapper[4813]: I1206 16:32:04.025321 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c426e12-23d7-4821-bdc3-882990a9f41e-utilities\") pod \"5c426e12-23d7-4821-bdc3-882990a9f41e\" (UID: \"5c426e12-23d7-4821-bdc3-882990a9f41e\") " Dec 06 16:32:04 crc kubenswrapper[4813]: I1206 16:32:04.025810 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c426e12-23d7-4821-bdc3-882990a9f41e-utilities" (OuterVolumeSpecName: "utilities") pod "5c426e12-23d7-4821-bdc3-882990a9f41e" (UID: "5c426e12-23d7-4821-bdc3-882990a9f41e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:32:04 crc kubenswrapper[4813]: I1206 16:32:04.025874 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62ssb\" (UniqueName: \"kubernetes.io/projected/5c426e12-23d7-4821-bdc3-882990a9f41e-kube-api-access-62ssb\") pod \"5c426e12-23d7-4821-bdc3-882990a9f41e\" (UID: \"5c426e12-23d7-4821-bdc3-882990a9f41e\") " Dec 06 16:32:04 crc kubenswrapper[4813]: I1206 16:32:04.026954 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c426e12-23d7-4821-bdc3-882990a9f41e-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 16:32:04 crc kubenswrapper[4813]: I1206 16:32:04.031468 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c426e12-23d7-4821-bdc3-882990a9f41e-kube-api-access-62ssb" (OuterVolumeSpecName: "kube-api-access-62ssb") pod "5c426e12-23d7-4821-bdc3-882990a9f41e" (UID: "5c426e12-23d7-4821-bdc3-882990a9f41e"). InnerVolumeSpecName "kube-api-access-62ssb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:32:04 crc kubenswrapper[4813]: I1206 16:32:04.069040 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c426e12-23d7-4821-bdc3-882990a9f41e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5c426e12-23d7-4821-bdc3-882990a9f41e" (UID: "5c426e12-23d7-4821-bdc3-882990a9f41e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:32:04 crc kubenswrapper[4813]: I1206 16:32:04.128938 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c426e12-23d7-4821-bdc3-882990a9f41e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 16:32:04 crc kubenswrapper[4813]: I1206 16:32:04.128969 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62ssb\" (UniqueName: \"kubernetes.io/projected/5c426e12-23d7-4821-bdc3-882990a9f41e-kube-api-access-62ssb\") on node \"crc\" DevicePath \"\"" Dec 06 16:32:04 crc kubenswrapper[4813]: I1206 16:32:04.351959 4813 generic.go:334] "Generic (PLEG): container finished" podID="5c426e12-23d7-4821-bdc3-882990a9f41e" containerID="497055b9b13dcdaea2551b9a0cc1cc221e0750445e1d6df16ba7a5ccc060f104" exitCode=0 Dec 06 16:32:04 crc kubenswrapper[4813]: I1206 16:32:04.352019 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7452r" event={"ID":"5c426e12-23d7-4821-bdc3-882990a9f41e","Type":"ContainerDied","Data":"497055b9b13dcdaea2551b9a0cc1cc221e0750445e1d6df16ba7a5ccc060f104"} Dec 06 16:32:04 crc kubenswrapper[4813]: I1206 16:32:04.352045 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7452r" Dec 06 16:32:04 crc kubenswrapper[4813]: I1206 16:32:04.352077 4813 scope.go:117] "RemoveContainer" containerID="497055b9b13dcdaea2551b9a0cc1cc221e0750445e1d6df16ba7a5ccc060f104" Dec 06 16:32:04 crc kubenswrapper[4813]: I1206 16:32:04.352059 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7452r" event={"ID":"5c426e12-23d7-4821-bdc3-882990a9f41e","Type":"ContainerDied","Data":"5913921de8c35f0d18ebce5103adefe45d8f223b00e22b023ee485f02f2f1708"} Dec 06 16:32:04 crc kubenswrapper[4813]: I1206 16:32:04.409032 4813 scope.go:117] "RemoveContainer" containerID="30e5fc5cf460b1f91911f313f7c6ccc0935930b8cf79615f6393160c78b7f10a" Dec 06 16:32:04 crc kubenswrapper[4813]: I1206 16:32:04.414744 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7452r"] Dec 06 16:32:04 crc kubenswrapper[4813]: I1206 16:32:04.429613 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7452r"] Dec 06 16:32:04 crc kubenswrapper[4813]: I1206 16:32:04.445048 4813 scope.go:117] "RemoveContainer" containerID="75c955ce0d5a8456ca005ae2f363da4277e17d813b8bb007e254154828141ba9" Dec 06 16:32:04 crc kubenswrapper[4813]: I1206 16:32:04.509649 4813 scope.go:117] "RemoveContainer" containerID="497055b9b13dcdaea2551b9a0cc1cc221e0750445e1d6df16ba7a5ccc060f104" Dec 06 16:32:04 crc kubenswrapper[4813]: E1206 16:32:04.510423 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"497055b9b13dcdaea2551b9a0cc1cc221e0750445e1d6df16ba7a5ccc060f104\": container with ID starting with 497055b9b13dcdaea2551b9a0cc1cc221e0750445e1d6df16ba7a5ccc060f104 not found: ID does not exist" containerID="497055b9b13dcdaea2551b9a0cc1cc221e0750445e1d6df16ba7a5ccc060f104" Dec 06 16:32:04 crc kubenswrapper[4813]: I1206 16:32:04.510456 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"497055b9b13dcdaea2551b9a0cc1cc221e0750445e1d6df16ba7a5ccc060f104"} err="failed to get container status \"497055b9b13dcdaea2551b9a0cc1cc221e0750445e1d6df16ba7a5ccc060f104\": rpc error: code = NotFound desc = could not find container \"497055b9b13dcdaea2551b9a0cc1cc221e0750445e1d6df16ba7a5ccc060f104\": container with ID starting with 497055b9b13dcdaea2551b9a0cc1cc221e0750445e1d6df16ba7a5ccc060f104 not found: ID does not exist" Dec 06 16:32:04 crc kubenswrapper[4813]: I1206 16:32:04.510478 4813 scope.go:117] "RemoveContainer" containerID="30e5fc5cf460b1f91911f313f7c6ccc0935930b8cf79615f6393160c78b7f10a" Dec 06 16:32:04 crc kubenswrapper[4813]: E1206 16:32:04.510908 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30e5fc5cf460b1f91911f313f7c6ccc0935930b8cf79615f6393160c78b7f10a\": container with ID starting with 30e5fc5cf460b1f91911f313f7c6ccc0935930b8cf79615f6393160c78b7f10a not found: ID does not exist" containerID="30e5fc5cf460b1f91911f313f7c6ccc0935930b8cf79615f6393160c78b7f10a" Dec 06 16:32:04 crc kubenswrapper[4813]: I1206 16:32:04.510976 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30e5fc5cf460b1f91911f313f7c6ccc0935930b8cf79615f6393160c78b7f10a"} err="failed to get container status \"30e5fc5cf460b1f91911f313f7c6ccc0935930b8cf79615f6393160c78b7f10a\": rpc error: code = NotFound desc = could not find container \"30e5fc5cf460b1f91911f313f7c6ccc0935930b8cf79615f6393160c78b7f10a\": container with ID starting with 30e5fc5cf460b1f91911f313f7c6ccc0935930b8cf79615f6393160c78b7f10a not found: ID does not exist" Dec 06 16:32:04 crc kubenswrapper[4813]: I1206 16:32:04.511126 4813 scope.go:117] "RemoveContainer" containerID="75c955ce0d5a8456ca005ae2f363da4277e17d813b8bb007e254154828141ba9" Dec 06 16:32:04 crc kubenswrapper[4813]: I1206 16:32:04.512625 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c426e12-23d7-4821-bdc3-882990a9f41e" path="/var/lib/kubelet/pods/5c426e12-23d7-4821-bdc3-882990a9f41e/volumes" Dec 06 16:32:04 crc kubenswrapper[4813]: E1206 16:32:04.512700 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75c955ce0d5a8456ca005ae2f363da4277e17d813b8bb007e254154828141ba9\": container with ID starting with 75c955ce0d5a8456ca005ae2f363da4277e17d813b8bb007e254154828141ba9 not found: ID does not exist" containerID="75c955ce0d5a8456ca005ae2f363da4277e17d813b8bb007e254154828141ba9" Dec 06 16:32:04 crc kubenswrapper[4813]: I1206 16:32:04.512728 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75c955ce0d5a8456ca005ae2f363da4277e17d813b8bb007e254154828141ba9"} err="failed to get container status \"75c955ce0d5a8456ca005ae2f363da4277e17d813b8bb007e254154828141ba9\": rpc error: code = NotFound desc = could not find container \"75c955ce0d5a8456ca005ae2f363da4277e17d813b8bb007e254154828141ba9\": container with ID starting with 75c955ce0d5a8456ca005ae2f363da4277e17d813b8bb007e254154828141ba9 not found: ID does not exist" Dec 06 16:32:05 crc kubenswrapper[4813]: I1206 16:32:05.570542 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-q4tb6" Dec 06 16:32:05 crc kubenswrapper[4813]: I1206 16:32:05.570884 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-q4tb6" Dec 06 16:32:05 crc kubenswrapper[4813]: I1206 16:32:05.636539 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-q4tb6" Dec 06 16:32:06 crc kubenswrapper[4813]: I1206 16:32:06.435088 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-q4tb6" Dec 06 16:32:07 crc kubenswrapper[4813]: I1206 16:32:07.243545 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q4tb6"] Dec 06 16:32:08 crc kubenswrapper[4813]: I1206 16:32:08.397517 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-q4tb6" podUID="7ee2a239-1568-4b54-9196-af436b40399c" containerName="registry-server" containerID="cri-o://7f9c405e09265913f0f7f8457dec061406d0e6f55e78e0ddd608f016482ee218" gracePeriod=2 Dec 06 16:32:08 crc kubenswrapper[4813]: I1206 16:32:08.911422 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q4tb6" Dec 06 16:32:09 crc kubenswrapper[4813]: I1206 16:32:09.034756 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ee2a239-1568-4b54-9196-af436b40399c-utilities\") pod \"7ee2a239-1568-4b54-9196-af436b40399c\" (UID: \"7ee2a239-1568-4b54-9196-af436b40399c\") " Dec 06 16:32:09 crc kubenswrapper[4813]: I1206 16:32:09.034910 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ee2a239-1568-4b54-9196-af436b40399c-catalog-content\") pod \"7ee2a239-1568-4b54-9196-af436b40399c\" (UID: \"7ee2a239-1568-4b54-9196-af436b40399c\") " Dec 06 16:32:09 crc kubenswrapper[4813]: I1206 16:32:09.034947 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksgdw\" (UniqueName: \"kubernetes.io/projected/7ee2a239-1568-4b54-9196-af436b40399c-kube-api-access-ksgdw\") pod \"7ee2a239-1568-4b54-9196-af436b40399c\" (UID: \"7ee2a239-1568-4b54-9196-af436b40399c\") " Dec 06 16:32:09 crc kubenswrapper[4813]: I1206 16:32:09.035722 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ee2a239-1568-4b54-9196-af436b40399c-utilities" (OuterVolumeSpecName: "utilities") pod "7ee2a239-1568-4b54-9196-af436b40399c" (UID: "7ee2a239-1568-4b54-9196-af436b40399c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:32:09 crc kubenswrapper[4813]: I1206 16:32:09.040496 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ee2a239-1568-4b54-9196-af436b40399c-kube-api-access-ksgdw" (OuterVolumeSpecName: "kube-api-access-ksgdw") pod "7ee2a239-1568-4b54-9196-af436b40399c" (UID: "7ee2a239-1568-4b54-9196-af436b40399c"). InnerVolumeSpecName "kube-api-access-ksgdw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:32:09 crc kubenswrapper[4813]: I1206 16:32:09.051922 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ee2a239-1568-4b54-9196-af436b40399c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7ee2a239-1568-4b54-9196-af436b40399c" (UID: "7ee2a239-1568-4b54-9196-af436b40399c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:32:09 crc kubenswrapper[4813]: I1206 16:32:09.137248 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ee2a239-1568-4b54-9196-af436b40399c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 16:32:09 crc kubenswrapper[4813]: I1206 16:32:09.137296 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksgdw\" (UniqueName: \"kubernetes.io/projected/7ee2a239-1568-4b54-9196-af436b40399c-kube-api-access-ksgdw\") on node \"crc\" DevicePath \"\"" Dec 06 16:32:09 crc kubenswrapper[4813]: I1206 16:32:09.137310 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ee2a239-1568-4b54-9196-af436b40399c-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 16:32:09 crc kubenswrapper[4813]: I1206 16:32:09.412310 4813 generic.go:334] "Generic (PLEG): container finished" podID="7ee2a239-1568-4b54-9196-af436b40399c" containerID="7f9c405e09265913f0f7f8457dec061406d0e6f55e78e0ddd608f016482ee218" exitCode=0 Dec 06 16:32:09 crc kubenswrapper[4813]: I1206 16:32:09.412359 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4tb6" event={"ID":"7ee2a239-1568-4b54-9196-af436b40399c","Type":"ContainerDied","Data":"7f9c405e09265913f0f7f8457dec061406d0e6f55e78e0ddd608f016482ee218"} Dec 06 16:32:09 crc kubenswrapper[4813]: I1206 16:32:09.412392 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4tb6" event={"ID":"7ee2a239-1568-4b54-9196-af436b40399c","Type":"ContainerDied","Data":"b2f483a7b048cf16b42e9b6b4eb90f0120f12fe92fe0e56e477cef2193d8edf5"} Dec 06 16:32:09 crc kubenswrapper[4813]: I1206 16:32:09.412411 4813 scope.go:117] "RemoveContainer" containerID="7f9c405e09265913f0f7f8457dec061406d0e6f55e78e0ddd608f016482ee218" Dec 06 16:32:09 crc kubenswrapper[4813]: I1206 16:32:09.412473 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q4tb6" Dec 06 16:32:09 crc kubenswrapper[4813]: I1206 16:32:09.447128 4813 scope.go:117] "RemoveContainer" containerID="1b9b388415e6f3aae84e93974a436e888daa72487976c183f81111ed89f56863" Dec 06 16:32:09 crc kubenswrapper[4813]: I1206 16:32:09.465840 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q4tb6"] Dec 06 16:32:09 crc kubenswrapper[4813]: I1206 16:32:09.473167 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-q4tb6"] Dec 06 16:32:09 crc kubenswrapper[4813]: I1206 16:32:09.491690 4813 scope.go:117] "RemoveContainer" containerID="187b5398f8c5398e130ffeb4f4d7b132a3721e70243bad31b0e0d8b66adae2eb" Dec 06 16:32:09 crc kubenswrapper[4813]: I1206 16:32:09.520284 4813 scope.go:117] "RemoveContainer" containerID="7f9c405e09265913f0f7f8457dec061406d0e6f55e78e0ddd608f016482ee218" Dec 06 16:32:09 crc kubenswrapper[4813]: E1206 16:32:09.520871 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f9c405e09265913f0f7f8457dec061406d0e6f55e78e0ddd608f016482ee218\": container with ID starting with 7f9c405e09265913f0f7f8457dec061406d0e6f55e78e0ddd608f016482ee218 not found: ID does not exist" containerID="7f9c405e09265913f0f7f8457dec061406d0e6f55e78e0ddd608f016482ee218" Dec 06 16:32:09 crc kubenswrapper[4813]: I1206 16:32:09.520909 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f9c405e09265913f0f7f8457dec061406d0e6f55e78e0ddd608f016482ee218"} err="failed to get container status \"7f9c405e09265913f0f7f8457dec061406d0e6f55e78e0ddd608f016482ee218\": rpc error: code = NotFound desc = could not find container \"7f9c405e09265913f0f7f8457dec061406d0e6f55e78e0ddd608f016482ee218\": container with ID starting with 7f9c405e09265913f0f7f8457dec061406d0e6f55e78e0ddd608f016482ee218 not found: ID does not exist" Dec 06 16:32:09 crc kubenswrapper[4813]: I1206 16:32:09.520935 4813 scope.go:117] "RemoveContainer" containerID="1b9b388415e6f3aae84e93974a436e888daa72487976c183f81111ed89f56863" Dec 06 16:32:09 crc kubenswrapper[4813]: E1206 16:32:09.521284 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b9b388415e6f3aae84e93974a436e888daa72487976c183f81111ed89f56863\": container with ID starting with 1b9b388415e6f3aae84e93974a436e888daa72487976c183f81111ed89f56863 not found: ID does not exist" containerID="1b9b388415e6f3aae84e93974a436e888daa72487976c183f81111ed89f56863" Dec 06 16:32:09 crc kubenswrapper[4813]: I1206 16:32:09.521307 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b9b388415e6f3aae84e93974a436e888daa72487976c183f81111ed89f56863"} err="failed to get container status \"1b9b388415e6f3aae84e93974a436e888daa72487976c183f81111ed89f56863\": rpc error: code = NotFound desc = could not find container \"1b9b388415e6f3aae84e93974a436e888daa72487976c183f81111ed89f56863\": container with ID starting with 1b9b388415e6f3aae84e93974a436e888daa72487976c183f81111ed89f56863 not found: ID does not exist" Dec 06 16:32:09 crc kubenswrapper[4813]: I1206 16:32:09.521323 4813 scope.go:117] "RemoveContainer" containerID="187b5398f8c5398e130ffeb4f4d7b132a3721e70243bad31b0e0d8b66adae2eb" Dec 06 16:32:09 crc kubenswrapper[4813]: E1206 16:32:09.521643 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"187b5398f8c5398e130ffeb4f4d7b132a3721e70243bad31b0e0d8b66adae2eb\": container with ID starting with 187b5398f8c5398e130ffeb4f4d7b132a3721e70243bad31b0e0d8b66adae2eb not found: ID does not exist" containerID="187b5398f8c5398e130ffeb4f4d7b132a3721e70243bad31b0e0d8b66adae2eb" Dec 06 16:32:09 crc kubenswrapper[4813]: I1206 16:32:09.521659 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"187b5398f8c5398e130ffeb4f4d7b132a3721e70243bad31b0e0d8b66adae2eb"} err="failed to get container status \"187b5398f8c5398e130ffeb4f4d7b132a3721e70243bad31b0e0d8b66adae2eb\": rpc error: code = NotFound desc = could not find container \"187b5398f8c5398e130ffeb4f4d7b132a3721e70243bad31b0e0d8b66adae2eb\": container with ID starting with 187b5398f8c5398e130ffeb4f4d7b132a3721e70243bad31b0e0d8b66adae2eb not found: ID does not exist" Dec 06 16:32:10 crc kubenswrapper[4813]: I1206 16:32:10.499485 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ee2a239-1568-4b54-9196-af436b40399c" path="/var/lib/kubelet/pods/7ee2a239-1568-4b54-9196-af436b40399c/volumes" Dec 06 16:32:19 crc kubenswrapper[4813]: I1206 16:32:19.428104 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:32:19 crc kubenswrapper[4813]: I1206 16:32:19.428567 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:32:19 crc kubenswrapper[4813]: I1206 16:32:19.428609 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" Dec 06 16:32:19 crc kubenswrapper[4813]: I1206 16:32:19.429290 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0f04b80ef02ad2942e19db0f6e8b08273d91a189f92b54348561c653b70e0191"} pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 16:32:19 crc kubenswrapper[4813]: I1206 16:32:19.429341 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" containerID="cri-o://0f04b80ef02ad2942e19db0f6e8b08273d91a189f92b54348561c653b70e0191" gracePeriod=600 Dec 06 16:32:20 crc kubenswrapper[4813]: I1206 16:32:20.564586 4813 generic.go:334] "Generic (PLEG): container finished" podID="d88e8bae-c055-4c55-b548-f621ff96de06" containerID="0f04b80ef02ad2942e19db0f6e8b08273d91a189f92b54348561c653b70e0191" exitCode=0 Dec 06 16:32:20 crc kubenswrapper[4813]: I1206 16:32:20.565291 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerDied","Data":"0f04b80ef02ad2942e19db0f6e8b08273d91a189f92b54348561c653b70e0191"} Dec 06 16:32:20 crc kubenswrapper[4813]: I1206 16:32:20.565373 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerStarted","Data":"8441b376fae9405da00e04a7e75ce187d0f25185f6fc0f6fc4224d7561e835bd"} Dec 06 16:32:20 crc kubenswrapper[4813]: I1206 16:32:20.565403 4813 scope.go:117] "RemoveContainer" containerID="1b8cc98443ea491aa620632d54d48664f2e8cd65b0790f1013b24051a06625ca" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.244198 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 06 16:32:47 crc kubenswrapper[4813]: E1206 16:32:47.246035 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c426e12-23d7-4821-bdc3-882990a9f41e" containerName="registry-server" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.246054 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c426e12-23d7-4821-bdc3-882990a9f41e" containerName="registry-server" Dec 06 16:32:47 crc kubenswrapper[4813]: E1206 16:32:47.246074 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ee2a239-1568-4b54-9196-af436b40399c" containerName="registry-server" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.246082 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ee2a239-1568-4b54-9196-af436b40399c" containerName="registry-server" Dec 06 16:32:47 crc kubenswrapper[4813]: E1206 16:32:47.246124 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ee2a239-1568-4b54-9196-af436b40399c" containerName="extract-utilities" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.246136 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ee2a239-1568-4b54-9196-af436b40399c" containerName="extract-utilities" Dec 06 16:32:47 crc kubenswrapper[4813]: E1206 16:32:47.246152 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1ecf1f7-01c3-456c-8698-1353e0651c24" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.246162 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1ecf1f7-01c3-456c-8698-1353e0651c24" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 06 16:32:47 crc kubenswrapper[4813]: E1206 16:32:47.246207 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ee2a239-1568-4b54-9196-af436b40399c" containerName="extract-content" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.246217 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ee2a239-1568-4b54-9196-af436b40399c" containerName="extract-content" Dec 06 16:32:47 crc kubenswrapper[4813]: E1206 16:32:47.246243 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c426e12-23d7-4821-bdc3-882990a9f41e" containerName="extract-content" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.246279 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c426e12-23d7-4821-bdc3-882990a9f41e" containerName="extract-content" Dec 06 16:32:47 crc kubenswrapper[4813]: E1206 16:32:47.246300 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c426e12-23d7-4821-bdc3-882990a9f41e" containerName="extract-utilities" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.246310 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c426e12-23d7-4821-bdc3-882990a9f41e" containerName="extract-utilities" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.250277 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1ecf1f7-01c3-456c-8698-1353e0651c24" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.250322 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c426e12-23d7-4821-bdc3-882990a9f41e" containerName="registry-server" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.250349 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ee2a239-1568-4b54-9196-af436b40399c" containerName="registry-server" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.251224 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.257487 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.257806 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.258059 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.258363 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-pf9vq" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.264797 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.310800 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " pod="openstack/tempest-tests-tempest" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.310877 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " pod="openstack/tempest-tests-tempest" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.310919 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-config-data\") pod \"tempest-tests-tempest\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " pod="openstack/tempest-tests-tempest" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.412775 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " pod="openstack/tempest-tests-tempest" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.412844 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r2f8\" (UniqueName: \"kubernetes.io/projected/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-kube-api-access-4r2f8\") pod \"tempest-tests-tempest\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " pod="openstack/tempest-tests-tempest" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.412907 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " pod="openstack/tempest-tests-tempest" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.412976 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " pod="openstack/tempest-tests-tempest" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.413089 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " pod="openstack/tempest-tests-tempest" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.414060 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " pod="openstack/tempest-tests-tempest" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.414137 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " pod="openstack/tempest-tests-tempest" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.414168 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " pod="openstack/tempest-tests-tempest" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.415106 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " pod="openstack/tempest-tests-tempest" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.415133 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-config-data\") pod \"tempest-tests-tempest\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " pod="openstack/tempest-tests-tempest" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.416409 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-config-data\") pod \"tempest-tests-tempest\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " pod="openstack/tempest-tests-tempest" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.420071 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " pod="openstack/tempest-tests-tempest" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.517736 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " pod="openstack/tempest-tests-tempest" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.517786 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " pod="openstack/tempest-tests-tempest" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.517959 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " pod="openstack/tempest-tests-tempest" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.518016 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r2f8\" (UniqueName: \"kubernetes.io/projected/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-kube-api-access-4r2f8\") pod \"tempest-tests-tempest\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " pod="openstack/tempest-tests-tempest" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.518058 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " pod="openstack/tempest-tests-tempest" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.518094 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " pod="openstack/tempest-tests-tempest" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.518716 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " pod="openstack/tempest-tests-tempest" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.519492 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/tempest-tests-tempest" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.520310 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " pod="openstack/tempest-tests-tempest" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.524813 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " pod="openstack/tempest-tests-tempest" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.526050 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " pod="openstack/tempest-tests-tempest" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.550852 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r2f8\" (UniqueName: \"kubernetes.io/projected/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-kube-api-access-4r2f8\") pod \"tempest-tests-tempest\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " pod="openstack/tempest-tests-tempest" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.555719 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " pod="openstack/tempest-tests-tempest" Dec 06 16:32:47 crc kubenswrapper[4813]: I1206 16:32:47.587422 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 06 16:32:48 crc kubenswrapper[4813]: I1206 16:32:48.144550 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 06 16:32:48 crc kubenswrapper[4813]: I1206 16:32:48.879414 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"5e4bd588-0ab4-4132-a2e8-1c5ea769f352","Type":"ContainerStarted","Data":"ab4d91cfe784eed4de6b8ec0c6479b5dc2ff09789bbb376ca814714b97cdf70f"} Dec 06 16:33:26 crc kubenswrapper[4813]: I1206 16:33:26.286510 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-f9nfd" podUID="3edd7ca8-a3ba-46c5-96c8-77aa9be672b7" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.64:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 16:33:28 crc kubenswrapper[4813]: E1206 16:33:28.189029 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 06 16:33:28 crc kubenswrapper[4813]: E1206 16:33:28.190668 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4r2f8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(5e4bd588-0ab4-4132-a2e8-1c5ea769f352): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 16:33:28 crc kubenswrapper[4813]: E1206 16:33:28.192137 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="5e4bd588-0ab4-4132-a2e8-1c5ea769f352" Dec 06 16:33:28 crc kubenswrapper[4813]: E1206 16:33:28.353218 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="5e4bd588-0ab4-4132-a2e8-1c5ea769f352" Dec 06 16:33:40 crc kubenswrapper[4813]: I1206 16:33:40.953737 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 06 16:33:42 crc kubenswrapper[4813]: I1206 16:33:42.527394 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"5e4bd588-0ab4-4132-a2e8-1c5ea769f352","Type":"ContainerStarted","Data":"0e98fefe803d2c7052168deb7e49def355f8e08c92f615c9862f6c25c493dcef"} Dec 06 16:33:42 crc kubenswrapper[4813]: I1206 16:33:42.552301 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.7464457209999997 podStartE2EDuration="56.552279614s" podCreationTimestamp="2025-12-06 16:32:46 +0000 UTC" firstStartedPulling="2025-12-06 16:32:48.143854315 +0000 UTC m=+2808.034733891" lastFinishedPulling="2025-12-06 16:33:40.949688198 +0000 UTC m=+2860.840567784" observedRunningTime="2025-12-06 16:33:42.549028439 +0000 UTC m=+2862.439908015" watchObservedRunningTime="2025-12-06 16:33:42.552279614 +0000 UTC m=+2862.443159200" Dec 06 16:34:19 crc kubenswrapper[4813]: I1206 16:34:19.427711 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:34:19 crc kubenswrapper[4813]: I1206 16:34:19.430409 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:34:49 crc kubenswrapper[4813]: I1206 16:34:49.427704 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:34:49 crc kubenswrapper[4813]: I1206 16:34:49.428275 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:35:19 crc kubenswrapper[4813]: I1206 16:35:19.427233 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:35:19 crc kubenswrapper[4813]: I1206 16:35:19.427763 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:35:19 crc kubenswrapper[4813]: I1206 16:35:19.427807 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" Dec 06 16:35:19 crc kubenswrapper[4813]: I1206 16:35:19.428509 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8441b376fae9405da00e04a7e75ce187d0f25185f6fc0f6fc4224d7561e835bd"} pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 16:35:19 crc kubenswrapper[4813]: I1206 16:35:19.428561 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" containerID="cri-o://8441b376fae9405da00e04a7e75ce187d0f25185f6fc0f6fc4224d7561e835bd" gracePeriod=600 Dec 06 16:35:19 crc kubenswrapper[4813]: E1206 16:35:19.609427 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:35:20 crc kubenswrapper[4813]: I1206 16:35:20.493591 4813 generic.go:334] "Generic (PLEG): container finished" podID="d88e8bae-c055-4c55-b548-f621ff96de06" containerID="8441b376fae9405da00e04a7e75ce187d0f25185f6fc0f6fc4224d7561e835bd" exitCode=0 Dec 06 16:35:20 crc kubenswrapper[4813]: I1206 16:35:20.498950 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerDied","Data":"8441b376fae9405da00e04a7e75ce187d0f25185f6fc0f6fc4224d7561e835bd"} Dec 06 16:35:20 crc kubenswrapper[4813]: I1206 16:35:20.499241 4813 scope.go:117] "RemoveContainer" containerID="0f04b80ef02ad2942e19db0f6e8b08273d91a189f92b54348561c653b70e0191" Dec 06 16:35:20 crc kubenswrapper[4813]: I1206 16:35:20.499792 4813 scope.go:117] "RemoveContainer" containerID="8441b376fae9405da00e04a7e75ce187d0f25185f6fc0f6fc4224d7561e835bd" Dec 06 16:35:20 crc kubenswrapper[4813]: E1206 16:35:20.500056 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:35:31 crc kubenswrapper[4813]: I1206 16:35:31.487138 4813 scope.go:117] "RemoveContainer" containerID="8441b376fae9405da00e04a7e75ce187d0f25185f6fc0f6fc4224d7561e835bd" Dec 06 16:35:31 crc kubenswrapper[4813]: E1206 16:35:31.487813 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:35:44 crc kubenswrapper[4813]: I1206 16:35:44.487713 4813 scope.go:117] "RemoveContainer" containerID="8441b376fae9405da00e04a7e75ce187d0f25185f6fc0f6fc4224d7561e835bd" Dec 06 16:35:44 crc kubenswrapper[4813]: E1206 16:35:44.488753 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:35:48 crc kubenswrapper[4813]: I1206 16:35:48.436437 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-l9k2r"] Dec 06 16:35:48 crc kubenswrapper[4813]: I1206 16:35:48.439722 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l9k2r" Dec 06 16:35:48 crc kubenswrapper[4813]: I1206 16:35:48.454345 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l9k2r"] Dec 06 16:35:48 crc kubenswrapper[4813]: I1206 16:35:48.478008 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5hx6\" (UniqueName: \"kubernetes.io/projected/0b465c01-166e-4314-939c-d538fc694d5f-kube-api-access-f5hx6\") pod \"redhat-operators-l9k2r\" (UID: \"0b465c01-166e-4314-939c-d538fc694d5f\") " pod="openshift-marketplace/redhat-operators-l9k2r" Dec 06 16:35:48 crc kubenswrapper[4813]: I1206 16:35:48.478111 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b465c01-166e-4314-939c-d538fc694d5f-utilities\") pod \"redhat-operators-l9k2r\" (UID: \"0b465c01-166e-4314-939c-d538fc694d5f\") " pod="openshift-marketplace/redhat-operators-l9k2r" Dec 06 16:35:48 crc kubenswrapper[4813]: I1206 16:35:48.478185 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b465c01-166e-4314-939c-d538fc694d5f-catalog-content\") pod \"redhat-operators-l9k2r\" (UID: \"0b465c01-166e-4314-939c-d538fc694d5f\") " pod="openshift-marketplace/redhat-operators-l9k2r" Dec 06 16:35:48 crc kubenswrapper[4813]: I1206 16:35:48.579562 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b465c01-166e-4314-939c-d538fc694d5f-catalog-content\") pod \"redhat-operators-l9k2r\" (UID: \"0b465c01-166e-4314-939c-d538fc694d5f\") " pod="openshift-marketplace/redhat-operators-l9k2r" Dec 06 16:35:48 crc kubenswrapper[4813]: I1206 16:35:48.580001 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5hx6\" (UniqueName: \"kubernetes.io/projected/0b465c01-166e-4314-939c-d538fc694d5f-kube-api-access-f5hx6\") pod \"redhat-operators-l9k2r\" (UID: \"0b465c01-166e-4314-939c-d538fc694d5f\") " pod="openshift-marketplace/redhat-operators-l9k2r" Dec 06 16:35:48 crc kubenswrapper[4813]: I1206 16:35:48.580110 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b465c01-166e-4314-939c-d538fc694d5f-utilities\") pod \"redhat-operators-l9k2r\" (UID: \"0b465c01-166e-4314-939c-d538fc694d5f\") " pod="openshift-marketplace/redhat-operators-l9k2r" Dec 06 16:35:48 crc kubenswrapper[4813]: I1206 16:35:48.580125 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b465c01-166e-4314-939c-d538fc694d5f-catalog-content\") pod \"redhat-operators-l9k2r\" (UID: \"0b465c01-166e-4314-939c-d538fc694d5f\") " pod="openshift-marketplace/redhat-operators-l9k2r" Dec 06 16:35:48 crc kubenswrapper[4813]: I1206 16:35:48.581033 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b465c01-166e-4314-939c-d538fc694d5f-utilities\") pod \"redhat-operators-l9k2r\" (UID: \"0b465c01-166e-4314-939c-d538fc694d5f\") " pod="openshift-marketplace/redhat-operators-l9k2r" Dec 06 16:35:48 crc kubenswrapper[4813]: I1206 16:35:48.672529 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5hx6\" (UniqueName: \"kubernetes.io/projected/0b465c01-166e-4314-939c-d538fc694d5f-kube-api-access-f5hx6\") pod \"redhat-operators-l9k2r\" (UID: \"0b465c01-166e-4314-939c-d538fc694d5f\") " pod="openshift-marketplace/redhat-operators-l9k2r" Dec 06 16:35:48 crc kubenswrapper[4813]: I1206 16:35:48.763059 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l9k2r" Dec 06 16:35:49 crc kubenswrapper[4813]: I1206 16:35:49.370602 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l9k2r"] Dec 06 16:35:49 crc kubenswrapper[4813]: I1206 16:35:49.772139 4813 generic.go:334] "Generic (PLEG): container finished" podID="0b465c01-166e-4314-939c-d538fc694d5f" containerID="39e81491aa6ae20479e42f5cb5211ffc15639e59f6f9ed3426fdbe0e5e0b1fca" exitCode=0 Dec 06 16:35:49 crc kubenswrapper[4813]: I1206 16:35:49.772235 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9k2r" event={"ID":"0b465c01-166e-4314-939c-d538fc694d5f","Type":"ContainerDied","Data":"39e81491aa6ae20479e42f5cb5211ffc15639e59f6f9ed3426fdbe0e5e0b1fca"} Dec 06 16:35:49 crc kubenswrapper[4813]: I1206 16:35:49.772494 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9k2r" event={"ID":"0b465c01-166e-4314-939c-d538fc694d5f","Type":"ContainerStarted","Data":"0e4a97bb0b3e72adc52016d2cedd818ba5cb3f432d0a3c928ba2e7e91c86332c"} Dec 06 16:35:49 crc kubenswrapper[4813]: I1206 16:35:49.774199 4813 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 16:35:50 crc kubenswrapper[4813]: I1206 16:35:50.789163 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9k2r" event={"ID":"0b465c01-166e-4314-939c-d538fc694d5f","Type":"ContainerStarted","Data":"6b20bf472714725bc7f750cdaad3b3e390e5e4d04a29bb30a8dae3feb7594e9c"} Dec 06 16:35:55 crc kubenswrapper[4813]: I1206 16:35:55.834050 4813 generic.go:334] "Generic (PLEG): container finished" podID="0b465c01-166e-4314-939c-d538fc694d5f" containerID="6b20bf472714725bc7f750cdaad3b3e390e5e4d04a29bb30a8dae3feb7594e9c" exitCode=0 Dec 06 16:35:55 crc kubenswrapper[4813]: I1206 16:35:55.834138 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9k2r" event={"ID":"0b465c01-166e-4314-939c-d538fc694d5f","Type":"ContainerDied","Data":"6b20bf472714725bc7f750cdaad3b3e390e5e4d04a29bb30a8dae3feb7594e9c"} Dec 06 16:35:56 crc kubenswrapper[4813]: I1206 16:35:56.847546 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9k2r" event={"ID":"0b465c01-166e-4314-939c-d538fc694d5f","Type":"ContainerStarted","Data":"b9d6c8724150a3f24c659e56ddc2a7253a405408117582f94e54a86b6653cb8b"} Dec 06 16:35:56 crc kubenswrapper[4813]: I1206 16:35:56.866720 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-l9k2r" podStartSLOduration=2.38822839 podStartE2EDuration="8.86670303s" podCreationTimestamp="2025-12-06 16:35:48 +0000 UTC" firstStartedPulling="2025-12-06 16:35:49.77389808 +0000 UTC m=+2989.664777656" lastFinishedPulling="2025-12-06 16:35:56.25237271 +0000 UTC m=+2996.143252296" observedRunningTime="2025-12-06 16:35:56.861660648 +0000 UTC m=+2996.752540224" watchObservedRunningTime="2025-12-06 16:35:56.86670303 +0000 UTC m=+2996.757582606" Dec 06 16:35:58 crc kubenswrapper[4813]: I1206 16:35:58.763426 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-l9k2r" Dec 06 16:35:58 crc kubenswrapper[4813]: I1206 16:35:58.764658 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-l9k2r" Dec 06 16:35:59 crc kubenswrapper[4813]: I1206 16:35:59.487455 4813 scope.go:117] "RemoveContainer" containerID="8441b376fae9405da00e04a7e75ce187d0f25185f6fc0f6fc4224d7561e835bd" Dec 06 16:35:59 crc kubenswrapper[4813]: E1206 16:35:59.487681 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:35:59 crc kubenswrapper[4813]: I1206 16:35:59.836401 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-l9k2r" podUID="0b465c01-166e-4314-939c-d538fc694d5f" containerName="registry-server" probeResult="failure" output=< Dec 06 16:35:59 crc kubenswrapper[4813]: timeout: failed to connect service ":50051" within 1s Dec 06 16:35:59 crc kubenswrapper[4813]: > Dec 06 16:36:08 crc kubenswrapper[4813]: I1206 16:36:08.819743 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-l9k2r" Dec 06 16:36:08 crc kubenswrapper[4813]: I1206 16:36:08.875812 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-l9k2r" Dec 06 16:36:09 crc kubenswrapper[4813]: I1206 16:36:09.064961 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l9k2r"] Dec 06 16:36:09 crc kubenswrapper[4813]: I1206 16:36:09.960226 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-l9k2r" podUID="0b465c01-166e-4314-939c-d538fc694d5f" containerName="registry-server" containerID="cri-o://b9d6c8724150a3f24c659e56ddc2a7253a405408117582f94e54a86b6653cb8b" gracePeriod=2 Dec 06 16:36:10 crc kubenswrapper[4813]: I1206 16:36:10.497355 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l9k2r" Dec 06 16:36:10 crc kubenswrapper[4813]: I1206 16:36:10.603597 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5hx6\" (UniqueName: \"kubernetes.io/projected/0b465c01-166e-4314-939c-d538fc694d5f-kube-api-access-f5hx6\") pod \"0b465c01-166e-4314-939c-d538fc694d5f\" (UID: \"0b465c01-166e-4314-939c-d538fc694d5f\") " Dec 06 16:36:10 crc kubenswrapper[4813]: I1206 16:36:10.604880 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b465c01-166e-4314-939c-d538fc694d5f-catalog-content\") pod \"0b465c01-166e-4314-939c-d538fc694d5f\" (UID: \"0b465c01-166e-4314-939c-d538fc694d5f\") " Dec 06 16:36:10 crc kubenswrapper[4813]: I1206 16:36:10.605066 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b465c01-166e-4314-939c-d538fc694d5f-utilities\") pod \"0b465c01-166e-4314-939c-d538fc694d5f\" (UID: \"0b465c01-166e-4314-939c-d538fc694d5f\") " Dec 06 16:36:10 crc kubenswrapper[4813]: I1206 16:36:10.605896 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b465c01-166e-4314-939c-d538fc694d5f-utilities" (OuterVolumeSpecName: "utilities") pod "0b465c01-166e-4314-939c-d538fc694d5f" (UID: "0b465c01-166e-4314-939c-d538fc694d5f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:36:10 crc kubenswrapper[4813]: I1206 16:36:10.611451 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b465c01-166e-4314-939c-d538fc694d5f-kube-api-access-f5hx6" (OuterVolumeSpecName: "kube-api-access-f5hx6") pod "0b465c01-166e-4314-939c-d538fc694d5f" (UID: "0b465c01-166e-4314-939c-d538fc694d5f"). InnerVolumeSpecName "kube-api-access-f5hx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:36:10 crc kubenswrapper[4813]: I1206 16:36:10.726463 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b465c01-166e-4314-939c-d538fc694d5f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0b465c01-166e-4314-939c-d538fc694d5f" (UID: "0b465c01-166e-4314-939c-d538fc694d5f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:36:10 crc kubenswrapper[4813]: I1206 16:36:10.745546 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b465c01-166e-4314-939c-d538fc694d5f-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 16:36:10 crc kubenswrapper[4813]: I1206 16:36:10.745575 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5hx6\" (UniqueName: \"kubernetes.io/projected/0b465c01-166e-4314-939c-d538fc694d5f-kube-api-access-f5hx6\") on node \"crc\" DevicePath \"\"" Dec 06 16:36:10 crc kubenswrapper[4813]: I1206 16:36:10.745585 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b465c01-166e-4314-939c-d538fc694d5f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 16:36:10 crc kubenswrapper[4813]: I1206 16:36:10.973026 4813 generic.go:334] "Generic (PLEG): container finished" podID="0b465c01-166e-4314-939c-d538fc694d5f" containerID="b9d6c8724150a3f24c659e56ddc2a7253a405408117582f94e54a86b6653cb8b" exitCode=0 Dec 06 16:36:10 crc kubenswrapper[4813]: I1206 16:36:10.973122 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l9k2r" Dec 06 16:36:10 crc kubenswrapper[4813]: I1206 16:36:10.973153 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9k2r" event={"ID":"0b465c01-166e-4314-939c-d538fc694d5f","Type":"ContainerDied","Data":"b9d6c8724150a3f24c659e56ddc2a7253a405408117582f94e54a86b6653cb8b"} Dec 06 16:36:10 crc kubenswrapper[4813]: I1206 16:36:10.973198 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9k2r" event={"ID":"0b465c01-166e-4314-939c-d538fc694d5f","Type":"ContainerDied","Data":"0e4a97bb0b3e72adc52016d2cedd818ba5cb3f432d0a3c928ba2e7e91c86332c"} Dec 06 16:36:10 crc kubenswrapper[4813]: I1206 16:36:10.973356 4813 scope.go:117] "RemoveContainer" containerID="b9d6c8724150a3f24c659e56ddc2a7253a405408117582f94e54a86b6653cb8b" Dec 06 16:36:11 crc kubenswrapper[4813]: I1206 16:36:11.009485 4813 scope.go:117] "RemoveContainer" containerID="6b20bf472714725bc7f750cdaad3b3e390e5e4d04a29bb30a8dae3feb7594e9c" Dec 06 16:36:11 crc kubenswrapper[4813]: I1206 16:36:11.035720 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l9k2r"] Dec 06 16:36:11 crc kubenswrapper[4813]: I1206 16:36:11.055305 4813 scope.go:117] "RemoveContainer" containerID="39e81491aa6ae20479e42f5cb5211ffc15639e59f6f9ed3426fdbe0e5e0b1fca" Dec 06 16:36:11 crc kubenswrapper[4813]: I1206 16:36:11.056228 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-l9k2r"] Dec 06 16:36:11 crc kubenswrapper[4813]: I1206 16:36:11.108492 4813 scope.go:117] "RemoveContainer" containerID="b9d6c8724150a3f24c659e56ddc2a7253a405408117582f94e54a86b6653cb8b" Dec 06 16:36:11 crc kubenswrapper[4813]: E1206 16:36:11.108938 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9d6c8724150a3f24c659e56ddc2a7253a405408117582f94e54a86b6653cb8b\": container with ID starting with b9d6c8724150a3f24c659e56ddc2a7253a405408117582f94e54a86b6653cb8b not found: ID does not exist" containerID="b9d6c8724150a3f24c659e56ddc2a7253a405408117582f94e54a86b6653cb8b" Dec 06 16:36:11 crc kubenswrapper[4813]: I1206 16:36:11.108970 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9d6c8724150a3f24c659e56ddc2a7253a405408117582f94e54a86b6653cb8b"} err="failed to get container status \"b9d6c8724150a3f24c659e56ddc2a7253a405408117582f94e54a86b6653cb8b\": rpc error: code = NotFound desc = could not find container \"b9d6c8724150a3f24c659e56ddc2a7253a405408117582f94e54a86b6653cb8b\": container with ID starting with b9d6c8724150a3f24c659e56ddc2a7253a405408117582f94e54a86b6653cb8b not found: ID does not exist" Dec 06 16:36:11 crc kubenswrapper[4813]: I1206 16:36:11.108991 4813 scope.go:117] "RemoveContainer" containerID="6b20bf472714725bc7f750cdaad3b3e390e5e4d04a29bb30a8dae3feb7594e9c" Dec 06 16:36:11 crc kubenswrapper[4813]: E1206 16:36:11.109184 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b20bf472714725bc7f750cdaad3b3e390e5e4d04a29bb30a8dae3feb7594e9c\": container with ID starting with 6b20bf472714725bc7f750cdaad3b3e390e5e4d04a29bb30a8dae3feb7594e9c not found: ID does not exist" containerID="6b20bf472714725bc7f750cdaad3b3e390e5e4d04a29bb30a8dae3feb7594e9c" Dec 06 16:36:11 crc kubenswrapper[4813]: I1206 16:36:11.109206 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b20bf472714725bc7f750cdaad3b3e390e5e4d04a29bb30a8dae3feb7594e9c"} err="failed to get container status \"6b20bf472714725bc7f750cdaad3b3e390e5e4d04a29bb30a8dae3feb7594e9c\": rpc error: code = NotFound desc = could not find container \"6b20bf472714725bc7f750cdaad3b3e390e5e4d04a29bb30a8dae3feb7594e9c\": container with ID starting with 6b20bf472714725bc7f750cdaad3b3e390e5e4d04a29bb30a8dae3feb7594e9c not found: ID does not exist" Dec 06 16:36:11 crc kubenswrapper[4813]: I1206 16:36:11.109221 4813 scope.go:117] "RemoveContainer" containerID="39e81491aa6ae20479e42f5cb5211ffc15639e59f6f9ed3426fdbe0e5e0b1fca" Dec 06 16:36:11 crc kubenswrapper[4813]: E1206 16:36:11.109793 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39e81491aa6ae20479e42f5cb5211ffc15639e59f6f9ed3426fdbe0e5e0b1fca\": container with ID starting with 39e81491aa6ae20479e42f5cb5211ffc15639e59f6f9ed3426fdbe0e5e0b1fca not found: ID does not exist" containerID="39e81491aa6ae20479e42f5cb5211ffc15639e59f6f9ed3426fdbe0e5e0b1fca" Dec 06 16:36:11 crc kubenswrapper[4813]: I1206 16:36:11.109820 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39e81491aa6ae20479e42f5cb5211ffc15639e59f6f9ed3426fdbe0e5e0b1fca"} err="failed to get container status \"39e81491aa6ae20479e42f5cb5211ffc15639e59f6f9ed3426fdbe0e5e0b1fca\": rpc error: code = NotFound desc = could not find container \"39e81491aa6ae20479e42f5cb5211ffc15639e59f6f9ed3426fdbe0e5e0b1fca\": container with ID starting with 39e81491aa6ae20479e42f5cb5211ffc15639e59f6f9ed3426fdbe0e5e0b1fca not found: ID does not exist" Dec 06 16:36:11 crc kubenswrapper[4813]: I1206 16:36:11.487229 4813 scope.go:117] "RemoveContainer" containerID="8441b376fae9405da00e04a7e75ce187d0f25185f6fc0f6fc4224d7561e835bd" Dec 06 16:36:11 crc kubenswrapper[4813]: E1206 16:36:11.487753 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:36:12 crc kubenswrapper[4813]: I1206 16:36:12.496711 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b465c01-166e-4314-939c-d538fc694d5f" path="/var/lib/kubelet/pods/0b465c01-166e-4314-939c-d538fc694d5f/volumes" Dec 06 16:36:26 crc kubenswrapper[4813]: I1206 16:36:26.487784 4813 scope.go:117] "RemoveContainer" containerID="8441b376fae9405da00e04a7e75ce187d0f25185f6fc0f6fc4224d7561e835bd" Dec 06 16:36:26 crc kubenswrapper[4813]: E1206 16:36:26.488493 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:36:39 crc kubenswrapper[4813]: I1206 16:36:39.488408 4813 scope.go:117] "RemoveContainer" containerID="8441b376fae9405da00e04a7e75ce187d0f25185f6fc0f6fc4224d7561e835bd" Dec 06 16:36:39 crc kubenswrapper[4813]: E1206 16:36:39.490398 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:36:54 crc kubenswrapper[4813]: I1206 16:36:54.487032 4813 scope.go:117] "RemoveContainer" containerID="8441b376fae9405da00e04a7e75ce187d0f25185f6fc0f6fc4224d7561e835bd" Dec 06 16:36:54 crc kubenswrapper[4813]: E1206 16:36:54.487755 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:37:09 crc kubenswrapper[4813]: I1206 16:37:09.487115 4813 scope.go:117] "RemoveContainer" containerID="8441b376fae9405da00e04a7e75ce187d0f25185f6fc0f6fc4224d7561e835bd" Dec 06 16:37:09 crc kubenswrapper[4813]: E1206 16:37:09.488957 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:37:24 crc kubenswrapper[4813]: I1206 16:37:24.487563 4813 scope.go:117] "RemoveContainer" containerID="8441b376fae9405da00e04a7e75ce187d0f25185f6fc0f6fc4224d7561e835bd" Dec 06 16:37:24 crc kubenswrapper[4813]: E1206 16:37:24.488556 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:37:36 crc kubenswrapper[4813]: I1206 16:37:36.487333 4813 scope.go:117] "RemoveContainer" containerID="8441b376fae9405da00e04a7e75ce187d0f25185f6fc0f6fc4224d7561e835bd" Dec 06 16:37:36 crc kubenswrapper[4813]: E1206 16:37:36.488042 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:37:47 crc kubenswrapper[4813]: I1206 16:37:47.487534 4813 scope.go:117] "RemoveContainer" containerID="8441b376fae9405da00e04a7e75ce187d0f25185f6fc0f6fc4224d7561e835bd" Dec 06 16:37:47 crc kubenswrapper[4813]: E1206 16:37:47.488396 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:38:01 crc kubenswrapper[4813]: I1206 16:38:01.488624 4813 scope.go:117] "RemoveContainer" containerID="8441b376fae9405da00e04a7e75ce187d0f25185f6fc0f6fc4224d7561e835bd" Dec 06 16:38:01 crc kubenswrapper[4813]: E1206 16:38:01.491846 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:38:13 crc kubenswrapper[4813]: I1206 16:38:13.487382 4813 scope.go:117] "RemoveContainer" containerID="8441b376fae9405da00e04a7e75ce187d0f25185f6fc0f6fc4224d7561e835bd" Dec 06 16:38:13 crc kubenswrapper[4813]: E1206 16:38:13.487991 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:38:27 crc kubenswrapper[4813]: I1206 16:38:27.486870 4813 scope.go:117] "RemoveContainer" containerID="8441b376fae9405da00e04a7e75ce187d0f25185f6fc0f6fc4224d7561e835bd" Dec 06 16:38:27 crc kubenswrapper[4813]: E1206 16:38:27.487463 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:38:41 crc kubenswrapper[4813]: I1206 16:38:41.486745 4813 scope.go:117] "RemoveContainer" containerID="8441b376fae9405da00e04a7e75ce187d0f25185f6fc0f6fc4224d7561e835bd" Dec 06 16:38:41 crc kubenswrapper[4813]: E1206 16:38:41.487704 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:38:54 crc kubenswrapper[4813]: I1206 16:38:54.487973 4813 scope.go:117] "RemoveContainer" containerID="8441b376fae9405da00e04a7e75ce187d0f25185f6fc0f6fc4224d7561e835bd" Dec 06 16:38:54 crc kubenswrapper[4813]: E1206 16:38:54.488949 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:38:59 crc kubenswrapper[4813]: I1206 16:38:59.612856 4813 generic.go:334] "Generic (PLEG): container finished" podID="5e4bd588-0ab4-4132-a2e8-1c5ea769f352" containerID="0e98fefe803d2c7052168deb7e49def355f8e08c92f615c9862f6c25c493dcef" exitCode=0 Dec 06 16:38:59 crc kubenswrapper[4813]: I1206 16:38:59.612985 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"5e4bd588-0ab4-4132-a2e8-1c5ea769f352","Type":"ContainerDied","Data":"0e98fefe803d2c7052168deb7e49def355f8e08c92f615c9862f6c25c493dcef"} Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.000350 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.030817 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-test-operator-ephemeral-temporary\") pod \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.030891 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.030980 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-config-data\") pod \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.031034 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-openstack-config-secret\") pod \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.031130 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4r2f8\" (UniqueName: \"kubernetes.io/projected/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-kube-api-access-4r2f8\") pod \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.031178 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-openstack-config\") pod \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.031251 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-ssh-key\") pod \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.031319 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-ca-certs\") pod \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.031388 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-test-operator-ephemeral-workdir\") pod \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\" (UID: \"5e4bd588-0ab4-4132-a2e8-1c5ea769f352\") " Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.031485 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "5e4bd588-0ab4-4132-a2e8-1c5ea769f352" (UID: "5e4bd588-0ab4-4132-a2e8-1c5ea769f352"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.031726 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-config-data" (OuterVolumeSpecName: "config-data") pod "5e4bd588-0ab4-4132-a2e8-1c5ea769f352" (UID: "5e4bd588-0ab4-4132-a2e8-1c5ea769f352"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.031855 4813 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.031877 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.037116 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-kube-api-access-4r2f8" (OuterVolumeSpecName: "kube-api-access-4r2f8") pod "5e4bd588-0ab4-4132-a2e8-1c5ea769f352" (UID: "5e4bd588-0ab4-4132-a2e8-1c5ea769f352"). InnerVolumeSpecName "kube-api-access-4r2f8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.052510 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "5e4bd588-0ab4-4132-a2e8-1c5ea769f352" (UID: "5e4bd588-0ab4-4132-a2e8-1c5ea769f352"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.056205 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "test-operator-logs") pod "5e4bd588-0ab4-4132-a2e8-1c5ea769f352" (UID: "5e4bd588-0ab4-4132-a2e8-1c5ea769f352"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.083781 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5e4bd588-0ab4-4132-a2e8-1c5ea769f352" (UID: "5e4bd588-0ab4-4132-a2e8-1c5ea769f352"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.086419 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "5e4bd588-0ab4-4132-a2e8-1c5ea769f352" (UID: "5e4bd588-0ab4-4132-a2e8-1c5ea769f352"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.089249 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "5e4bd588-0ab4-4132-a2e8-1c5ea769f352" (UID: "5e4bd588-0ab4-4132-a2e8-1c5ea769f352"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.111877 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "5e4bd588-0ab4-4132-a2e8-1c5ea769f352" (UID: "5e4bd588-0ab4-4132-a2e8-1c5ea769f352"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.133954 4813 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.133995 4813 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.134033 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.134046 4813 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.134058 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4r2f8\" (UniqueName: \"kubernetes.io/projected/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-kube-api-access-4r2f8\") on node \"crc\" DevicePath \"\"" Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.134069 4813 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.134078 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5e4bd588-0ab4-4132-a2e8-1c5ea769f352-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.152854 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.235928 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.639189 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"5e4bd588-0ab4-4132-a2e8-1c5ea769f352","Type":"ContainerDied","Data":"ab4d91cfe784eed4de6b8ec0c6479b5dc2ff09789bbb376ca814714b97cdf70f"} Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.639561 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab4d91cfe784eed4de6b8ec0c6479b5dc2ff09789bbb376ca814714b97cdf70f" Dec 06 16:39:01 crc kubenswrapper[4813]: I1206 16:39:01.639626 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 06 16:39:05 crc kubenswrapper[4813]: I1206 16:39:05.487528 4813 scope.go:117] "RemoveContainer" containerID="8441b376fae9405da00e04a7e75ce187d0f25185f6fc0f6fc4224d7561e835bd" Dec 06 16:39:05 crc kubenswrapper[4813]: E1206 16:39:05.488296 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:39:06 crc kubenswrapper[4813]: I1206 16:39:06.102072 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 06 16:39:06 crc kubenswrapper[4813]: E1206 16:39:06.102862 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b465c01-166e-4314-939c-d538fc694d5f" containerName="extract-content" Dec 06 16:39:06 crc kubenswrapper[4813]: I1206 16:39:06.102894 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b465c01-166e-4314-939c-d538fc694d5f" containerName="extract-content" Dec 06 16:39:06 crc kubenswrapper[4813]: E1206 16:39:06.102946 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b465c01-166e-4314-939c-d538fc694d5f" containerName="extract-utilities" Dec 06 16:39:06 crc kubenswrapper[4813]: I1206 16:39:06.102971 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b465c01-166e-4314-939c-d538fc694d5f" containerName="extract-utilities" Dec 06 16:39:06 crc kubenswrapper[4813]: E1206 16:39:06.102998 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b465c01-166e-4314-939c-d538fc694d5f" containerName="registry-server" Dec 06 16:39:06 crc kubenswrapper[4813]: I1206 16:39:06.103011 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b465c01-166e-4314-939c-d538fc694d5f" containerName="registry-server" Dec 06 16:39:06 crc kubenswrapper[4813]: E1206 16:39:06.103042 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e4bd588-0ab4-4132-a2e8-1c5ea769f352" containerName="tempest-tests-tempest-tests-runner" Dec 06 16:39:06 crc kubenswrapper[4813]: I1206 16:39:06.103055 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e4bd588-0ab4-4132-a2e8-1c5ea769f352" containerName="tempest-tests-tempest-tests-runner" Dec 06 16:39:06 crc kubenswrapper[4813]: I1206 16:39:06.103444 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e4bd588-0ab4-4132-a2e8-1c5ea769f352" containerName="tempest-tests-tempest-tests-runner" Dec 06 16:39:06 crc kubenswrapper[4813]: I1206 16:39:06.103476 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b465c01-166e-4314-939c-d538fc694d5f" containerName="registry-server" Dec 06 16:39:06 crc kubenswrapper[4813]: I1206 16:39:06.104499 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 16:39:06 crc kubenswrapper[4813]: I1206 16:39:06.108409 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-pf9vq" Dec 06 16:39:06 crc kubenswrapper[4813]: I1206 16:39:06.117707 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 06 16:39:06 crc kubenswrapper[4813]: I1206 16:39:06.154501 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"9899c9cb-fe20-41a1-a200-f70d5b63b314\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 16:39:06 crc kubenswrapper[4813]: I1206 16:39:06.154588 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5l4hb\" (UniqueName: \"kubernetes.io/projected/9899c9cb-fe20-41a1-a200-f70d5b63b314-kube-api-access-5l4hb\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"9899c9cb-fe20-41a1-a200-f70d5b63b314\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 16:39:06 crc kubenswrapper[4813]: I1206 16:39:06.256803 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"9899c9cb-fe20-41a1-a200-f70d5b63b314\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 16:39:06 crc kubenswrapper[4813]: I1206 16:39:06.256899 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5l4hb\" (UniqueName: \"kubernetes.io/projected/9899c9cb-fe20-41a1-a200-f70d5b63b314-kube-api-access-5l4hb\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"9899c9cb-fe20-41a1-a200-f70d5b63b314\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 16:39:06 crc kubenswrapper[4813]: I1206 16:39:06.257129 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"9899c9cb-fe20-41a1-a200-f70d5b63b314\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 16:39:06 crc kubenswrapper[4813]: I1206 16:39:06.288609 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5l4hb\" (UniqueName: \"kubernetes.io/projected/9899c9cb-fe20-41a1-a200-f70d5b63b314-kube-api-access-5l4hb\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"9899c9cb-fe20-41a1-a200-f70d5b63b314\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 16:39:06 crc kubenswrapper[4813]: I1206 16:39:06.329936 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"9899c9cb-fe20-41a1-a200-f70d5b63b314\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 16:39:06 crc kubenswrapper[4813]: I1206 16:39:06.441021 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 16:39:06 crc kubenswrapper[4813]: I1206 16:39:06.981835 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 06 16:39:07 crc kubenswrapper[4813]: I1206 16:39:07.713876 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"9899c9cb-fe20-41a1-a200-f70d5b63b314","Type":"ContainerStarted","Data":"accd84f7d1dab64a21ec4933ae2d58265d9fa4dc1f48c6683ed78d2c3834cdb3"} Dec 06 16:39:08 crc kubenswrapper[4813]: I1206 16:39:08.729379 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"9899c9cb-fe20-41a1-a200-f70d5b63b314","Type":"ContainerStarted","Data":"9b8906b89e8e6b8349387abf62e15821999d620de8f8e139c2b08aa6192c86c6"} Dec 06 16:39:08 crc kubenswrapper[4813]: I1206 16:39:08.746991 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.7130726269999998 podStartE2EDuration="2.746969809s" podCreationTimestamp="2025-12-06 16:39:06 +0000 UTC" firstStartedPulling="2025-12-06 16:39:06.985819441 +0000 UTC m=+3186.876699027" lastFinishedPulling="2025-12-06 16:39:08.019716623 +0000 UTC m=+3187.910596209" observedRunningTime="2025-12-06 16:39:08.745008667 +0000 UTC m=+3188.635888243" watchObservedRunningTime="2025-12-06 16:39:08.746969809 +0000 UTC m=+3188.637849385" Dec 06 16:39:16 crc kubenswrapper[4813]: I1206 16:39:16.487731 4813 scope.go:117] "RemoveContainer" containerID="8441b376fae9405da00e04a7e75ce187d0f25185f6fc0f6fc4224d7561e835bd" Dec 06 16:39:16 crc kubenswrapper[4813]: E1206 16:39:16.488960 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:39:31 crc kubenswrapper[4813]: I1206 16:39:31.487912 4813 scope.go:117] "RemoveContainer" containerID="8441b376fae9405da00e04a7e75ce187d0f25185f6fc0f6fc4224d7561e835bd" Dec 06 16:39:31 crc kubenswrapper[4813]: E1206 16:39:31.489076 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:39:32 crc kubenswrapper[4813]: I1206 16:39:32.222634 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jlgf8/must-gather-jzpxn"] Dec 06 16:39:32 crc kubenswrapper[4813]: I1206 16:39:32.224355 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jlgf8/must-gather-jzpxn" Dec 06 16:39:32 crc kubenswrapper[4813]: I1206 16:39:32.228548 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-jlgf8"/"kube-root-ca.crt" Dec 06 16:39:32 crc kubenswrapper[4813]: I1206 16:39:32.228808 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-jlgf8"/"openshift-service-ca.crt" Dec 06 16:39:32 crc kubenswrapper[4813]: I1206 16:39:32.266924 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-jlgf8/must-gather-jzpxn"] Dec 06 16:39:32 crc kubenswrapper[4813]: I1206 16:39:32.378275 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cj8w9\" (UniqueName: \"kubernetes.io/projected/3c20161a-c3b7-4359-882f-7764fc0908ca-kube-api-access-cj8w9\") pod \"must-gather-jzpxn\" (UID: \"3c20161a-c3b7-4359-882f-7764fc0908ca\") " pod="openshift-must-gather-jlgf8/must-gather-jzpxn" Dec 06 16:39:32 crc kubenswrapper[4813]: I1206 16:39:32.378346 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3c20161a-c3b7-4359-882f-7764fc0908ca-must-gather-output\") pod \"must-gather-jzpxn\" (UID: \"3c20161a-c3b7-4359-882f-7764fc0908ca\") " pod="openshift-must-gather-jlgf8/must-gather-jzpxn" Dec 06 16:39:32 crc kubenswrapper[4813]: I1206 16:39:32.480166 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3c20161a-c3b7-4359-882f-7764fc0908ca-must-gather-output\") pod \"must-gather-jzpxn\" (UID: \"3c20161a-c3b7-4359-882f-7764fc0908ca\") " pod="openshift-must-gather-jlgf8/must-gather-jzpxn" Dec 06 16:39:32 crc kubenswrapper[4813]: I1206 16:39:32.480417 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cj8w9\" (UniqueName: \"kubernetes.io/projected/3c20161a-c3b7-4359-882f-7764fc0908ca-kube-api-access-cj8w9\") pod \"must-gather-jzpxn\" (UID: \"3c20161a-c3b7-4359-882f-7764fc0908ca\") " pod="openshift-must-gather-jlgf8/must-gather-jzpxn" Dec 06 16:39:32 crc kubenswrapper[4813]: I1206 16:39:32.480780 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3c20161a-c3b7-4359-882f-7764fc0908ca-must-gather-output\") pod \"must-gather-jzpxn\" (UID: \"3c20161a-c3b7-4359-882f-7764fc0908ca\") " pod="openshift-must-gather-jlgf8/must-gather-jzpxn" Dec 06 16:39:32 crc kubenswrapper[4813]: I1206 16:39:32.499882 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cj8w9\" (UniqueName: \"kubernetes.io/projected/3c20161a-c3b7-4359-882f-7764fc0908ca-kube-api-access-cj8w9\") pod \"must-gather-jzpxn\" (UID: \"3c20161a-c3b7-4359-882f-7764fc0908ca\") " pod="openshift-must-gather-jlgf8/must-gather-jzpxn" Dec 06 16:39:32 crc kubenswrapper[4813]: I1206 16:39:32.547877 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jlgf8/must-gather-jzpxn" Dec 06 16:39:33 crc kubenswrapper[4813]: I1206 16:39:33.012314 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-jlgf8/must-gather-jzpxn"] Dec 06 16:39:34 crc kubenswrapper[4813]: I1206 16:39:34.033234 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jlgf8/must-gather-jzpxn" event={"ID":"3c20161a-c3b7-4359-882f-7764fc0908ca","Type":"ContainerStarted","Data":"f7bb75da50e5694f7386a4aa434cecfe3e6ab2d0639b2f228bca9abb2bc6bb17"} Dec 06 16:39:38 crc kubenswrapper[4813]: I1206 16:39:38.069010 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jlgf8/must-gather-jzpxn" event={"ID":"3c20161a-c3b7-4359-882f-7764fc0908ca","Type":"ContainerStarted","Data":"99b49e927f9e2d1e28a20400fe90995d8a0239783de2d407ed4c2825fbb52cf4"} Dec 06 16:39:38 crc kubenswrapper[4813]: I1206 16:39:38.069299 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jlgf8/must-gather-jzpxn" event={"ID":"3c20161a-c3b7-4359-882f-7764fc0908ca","Type":"ContainerStarted","Data":"8152527cee71164fbc9fb5ff72a2ac2f96023937781fcb9951718f1ff107942f"} Dec 06 16:39:38 crc kubenswrapper[4813]: I1206 16:39:38.088254 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-jlgf8/must-gather-jzpxn" podStartSLOduration=1.692836762 podStartE2EDuration="6.088241059s" podCreationTimestamp="2025-12-06 16:39:32 +0000 UTC" firstStartedPulling="2025-12-06 16:39:33.03063803 +0000 UTC m=+3212.921517606" lastFinishedPulling="2025-12-06 16:39:37.426042327 +0000 UTC m=+3217.316921903" observedRunningTime="2025-12-06 16:39:38.083168486 +0000 UTC m=+3217.974048062" watchObservedRunningTime="2025-12-06 16:39:38.088241059 +0000 UTC m=+3217.979120635" Dec 06 16:39:41 crc kubenswrapper[4813]: I1206 16:39:41.894648 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jlgf8/crc-debug-jnnkw"] Dec 06 16:39:41 crc kubenswrapper[4813]: I1206 16:39:41.896215 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jlgf8/crc-debug-jnnkw" Dec 06 16:39:41 crc kubenswrapper[4813]: I1206 16:39:41.899153 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-jlgf8"/"default-dockercfg-s4b7p" Dec 06 16:39:41 crc kubenswrapper[4813]: I1206 16:39:41.967212 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1e7de122-b529-4ee9-ac39-a31a22fc19e9-host\") pod \"crc-debug-jnnkw\" (UID: \"1e7de122-b529-4ee9-ac39-a31a22fc19e9\") " pod="openshift-must-gather-jlgf8/crc-debug-jnnkw" Dec 06 16:39:41 crc kubenswrapper[4813]: I1206 16:39:41.967425 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pwc4\" (UniqueName: \"kubernetes.io/projected/1e7de122-b529-4ee9-ac39-a31a22fc19e9-kube-api-access-2pwc4\") pod \"crc-debug-jnnkw\" (UID: \"1e7de122-b529-4ee9-ac39-a31a22fc19e9\") " pod="openshift-must-gather-jlgf8/crc-debug-jnnkw" Dec 06 16:39:42 crc kubenswrapper[4813]: I1206 16:39:42.069240 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1e7de122-b529-4ee9-ac39-a31a22fc19e9-host\") pod \"crc-debug-jnnkw\" (UID: \"1e7de122-b529-4ee9-ac39-a31a22fc19e9\") " pod="openshift-must-gather-jlgf8/crc-debug-jnnkw" Dec 06 16:39:42 crc kubenswrapper[4813]: I1206 16:39:42.069362 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1e7de122-b529-4ee9-ac39-a31a22fc19e9-host\") pod \"crc-debug-jnnkw\" (UID: \"1e7de122-b529-4ee9-ac39-a31a22fc19e9\") " pod="openshift-must-gather-jlgf8/crc-debug-jnnkw" Dec 06 16:39:42 crc kubenswrapper[4813]: I1206 16:39:42.069424 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pwc4\" (UniqueName: \"kubernetes.io/projected/1e7de122-b529-4ee9-ac39-a31a22fc19e9-kube-api-access-2pwc4\") pod \"crc-debug-jnnkw\" (UID: \"1e7de122-b529-4ee9-ac39-a31a22fc19e9\") " pod="openshift-must-gather-jlgf8/crc-debug-jnnkw" Dec 06 16:39:42 crc kubenswrapper[4813]: I1206 16:39:42.099806 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pwc4\" (UniqueName: \"kubernetes.io/projected/1e7de122-b529-4ee9-ac39-a31a22fc19e9-kube-api-access-2pwc4\") pod \"crc-debug-jnnkw\" (UID: \"1e7de122-b529-4ee9-ac39-a31a22fc19e9\") " pod="openshift-must-gather-jlgf8/crc-debug-jnnkw" Dec 06 16:39:42 crc kubenswrapper[4813]: I1206 16:39:42.217913 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jlgf8/crc-debug-jnnkw" Dec 06 16:39:42 crc kubenswrapper[4813]: W1206 16:39:42.247406 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e7de122_b529_4ee9_ac39_a31a22fc19e9.slice/crio-6acdddc5c66914f631103fa2e292512e875dc97133ca549e88f82e7e38832d83 WatchSource:0}: Error finding container 6acdddc5c66914f631103fa2e292512e875dc97133ca549e88f82e7e38832d83: Status 404 returned error can't find the container with id 6acdddc5c66914f631103fa2e292512e875dc97133ca549e88f82e7e38832d83 Dec 06 16:39:43 crc kubenswrapper[4813]: I1206 16:39:43.120502 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jlgf8/crc-debug-jnnkw" event={"ID":"1e7de122-b529-4ee9-ac39-a31a22fc19e9","Type":"ContainerStarted","Data":"6acdddc5c66914f631103fa2e292512e875dc97133ca549e88f82e7e38832d83"} Dec 06 16:39:45 crc kubenswrapper[4813]: I1206 16:39:45.487318 4813 scope.go:117] "RemoveContainer" containerID="8441b376fae9405da00e04a7e75ce187d0f25185f6fc0f6fc4224d7561e835bd" Dec 06 16:39:45 crc kubenswrapper[4813]: E1206 16:39:45.487902 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:39:54 crc kubenswrapper[4813]: I1206 16:39:54.224865 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jlgf8/crc-debug-jnnkw" event={"ID":"1e7de122-b529-4ee9-ac39-a31a22fc19e9","Type":"ContainerStarted","Data":"e8eeef2364d649555a536ddd16a7b6a11037a5c4688dae5922f1f23c4a727276"} Dec 06 16:39:54 crc kubenswrapper[4813]: I1206 16:39:54.245618 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-jlgf8/crc-debug-jnnkw" podStartSLOduration=1.859565616 podStartE2EDuration="13.245598649s" podCreationTimestamp="2025-12-06 16:39:41 +0000 UTC" firstStartedPulling="2025-12-06 16:39:42.249553279 +0000 UTC m=+3222.140432855" lastFinishedPulling="2025-12-06 16:39:53.635586312 +0000 UTC m=+3233.526465888" observedRunningTime="2025-12-06 16:39:54.239043727 +0000 UTC m=+3234.129923303" watchObservedRunningTime="2025-12-06 16:39:54.245598649 +0000 UTC m=+3234.136478225" Dec 06 16:39:57 crc kubenswrapper[4813]: I1206 16:39:57.486997 4813 scope.go:117] "RemoveContainer" containerID="8441b376fae9405da00e04a7e75ce187d0f25185f6fc0f6fc4224d7561e835bd" Dec 06 16:39:57 crc kubenswrapper[4813]: E1206 16:39:57.487653 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:40:08 crc kubenswrapper[4813]: I1206 16:40:08.432727 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-n8lrc"] Dec 06 16:40:08 crc kubenswrapper[4813]: I1206 16:40:08.439978 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n8lrc" Dec 06 16:40:08 crc kubenswrapper[4813]: I1206 16:40:08.442980 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n8lrc"] Dec 06 16:40:08 crc kubenswrapper[4813]: I1206 16:40:08.575584 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ad84257-a4c7-4674-9329-f862f8b1460e-catalog-content\") pod \"certified-operators-n8lrc\" (UID: \"2ad84257-a4c7-4674-9329-f862f8b1460e\") " pod="openshift-marketplace/certified-operators-n8lrc" Dec 06 16:40:08 crc kubenswrapper[4813]: I1206 16:40:08.576101 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ad84257-a4c7-4674-9329-f862f8b1460e-utilities\") pod \"certified-operators-n8lrc\" (UID: \"2ad84257-a4c7-4674-9329-f862f8b1460e\") " pod="openshift-marketplace/certified-operators-n8lrc" Dec 06 16:40:08 crc kubenswrapper[4813]: I1206 16:40:08.576178 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcl4k\" (UniqueName: \"kubernetes.io/projected/2ad84257-a4c7-4674-9329-f862f8b1460e-kube-api-access-zcl4k\") pod \"certified-operators-n8lrc\" (UID: \"2ad84257-a4c7-4674-9329-f862f8b1460e\") " pod="openshift-marketplace/certified-operators-n8lrc" Dec 06 16:40:08 crc kubenswrapper[4813]: I1206 16:40:08.678360 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ad84257-a4c7-4674-9329-f862f8b1460e-catalog-content\") pod \"certified-operators-n8lrc\" (UID: \"2ad84257-a4c7-4674-9329-f862f8b1460e\") " pod="openshift-marketplace/certified-operators-n8lrc" Dec 06 16:40:08 crc kubenswrapper[4813]: I1206 16:40:08.678508 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ad84257-a4c7-4674-9329-f862f8b1460e-utilities\") pod \"certified-operators-n8lrc\" (UID: \"2ad84257-a4c7-4674-9329-f862f8b1460e\") " pod="openshift-marketplace/certified-operators-n8lrc" Dec 06 16:40:08 crc kubenswrapper[4813]: I1206 16:40:08.678559 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcl4k\" (UniqueName: \"kubernetes.io/projected/2ad84257-a4c7-4674-9329-f862f8b1460e-kube-api-access-zcl4k\") pod \"certified-operators-n8lrc\" (UID: \"2ad84257-a4c7-4674-9329-f862f8b1460e\") " pod="openshift-marketplace/certified-operators-n8lrc" Dec 06 16:40:08 crc kubenswrapper[4813]: I1206 16:40:08.678913 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ad84257-a4c7-4674-9329-f862f8b1460e-catalog-content\") pod \"certified-operators-n8lrc\" (UID: \"2ad84257-a4c7-4674-9329-f862f8b1460e\") " pod="openshift-marketplace/certified-operators-n8lrc" Dec 06 16:40:08 crc kubenswrapper[4813]: I1206 16:40:08.679175 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ad84257-a4c7-4674-9329-f862f8b1460e-utilities\") pod \"certified-operators-n8lrc\" (UID: \"2ad84257-a4c7-4674-9329-f862f8b1460e\") " pod="openshift-marketplace/certified-operators-n8lrc" Dec 06 16:40:08 crc kubenswrapper[4813]: I1206 16:40:08.699724 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcl4k\" (UniqueName: \"kubernetes.io/projected/2ad84257-a4c7-4674-9329-f862f8b1460e-kube-api-access-zcl4k\") pod \"certified-operators-n8lrc\" (UID: \"2ad84257-a4c7-4674-9329-f862f8b1460e\") " pod="openshift-marketplace/certified-operators-n8lrc" Dec 06 16:40:08 crc kubenswrapper[4813]: I1206 16:40:08.764179 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n8lrc" Dec 06 16:40:09 crc kubenswrapper[4813]: I1206 16:40:09.332534 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n8lrc"] Dec 06 16:40:09 crc kubenswrapper[4813]: I1206 16:40:09.370465 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n8lrc" event={"ID":"2ad84257-a4c7-4674-9329-f862f8b1460e","Type":"ContainerStarted","Data":"af4f4297a2091b8ac9befccbf99a70901b97f430e25cdf94d97c5c955f31181d"} Dec 06 16:40:10 crc kubenswrapper[4813]: I1206 16:40:10.382277 4813 generic.go:334] "Generic (PLEG): container finished" podID="2ad84257-a4c7-4674-9329-f862f8b1460e" containerID="df271b90b410490a6c7188e83a8a51e95be03a0d4d3af97f29a3ba8971903d80" exitCode=0 Dec 06 16:40:10 crc kubenswrapper[4813]: I1206 16:40:10.382783 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n8lrc" event={"ID":"2ad84257-a4c7-4674-9329-f862f8b1460e","Type":"ContainerDied","Data":"df271b90b410490a6c7188e83a8a51e95be03a0d4d3af97f29a3ba8971903d80"} Dec 06 16:40:12 crc kubenswrapper[4813]: I1206 16:40:12.400711 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n8lrc" event={"ID":"2ad84257-a4c7-4674-9329-f862f8b1460e","Type":"ContainerStarted","Data":"902280a2744bfb274515b042117c93227dd198e40c5688b88f374f488e40068d"} Dec 06 16:40:12 crc kubenswrapper[4813]: I1206 16:40:12.489516 4813 scope.go:117] "RemoveContainer" containerID="8441b376fae9405da00e04a7e75ce187d0f25185f6fc0f6fc4224d7561e835bd" Dec 06 16:40:12 crc kubenswrapper[4813]: E1206 16:40:12.489780 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:40:13 crc kubenswrapper[4813]: I1206 16:40:13.410194 4813 generic.go:334] "Generic (PLEG): container finished" podID="2ad84257-a4c7-4674-9329-f862f8b1460e" containerID="902280a2744bfb274515b042117c93227dd198e40c5688b88f374f488e40068d" exitCode=0 Dec 06 16:40:13 crc kubenswrapper[4813]: I1206 16:40:13.410693 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n8lrc" event={"ID":"2ad84257-a4c7-4674-9329-f862f8b1460e","Type":"ContainerDied","Data":"902280a2744bfb274515b042117c93227dd198e40c5688b88f374f488e40068d"} Dec 06 16:40:14 crc kubenswrapper[4813]: I1206 16:40:14.420211 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n8lrc" event={"ID":"2ad84257-a4c7-4674-9329-f862f8b1460e","Type":"ContainerStarted","Data":"2ba915cf88622272e091db38700585f1a8b575a946d2da70b7efcf3983a91ea5"} Dec 06 16:40:14 crc kubenswrapper[4813]: I1206 16:40:14.442083 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-n8lrc" podStartSLOduration=3.012215206 podStartE2EDuration="6.442066105s" podCreationTimestamp="2025-12-06 16:40:08 +0000 UTC" firstStartedPulling="2025-12-06 16:40:10.385097736 +0000 UTC m=+3250.275977312" lastFinishedPulling="2025-12-06 16:40:13.814948635 +0000 UTC m=+3253.705828211" observedRunningTime="2025-12-06 16:40:14.43922976 +0000 UTC m=+3254.330109336" watchObservedRunningTime="2025-12-06 16:40:14.442066105 +0000 UTC m=+3254.332945681" Dec 06 16:40:18 crc kubenswrapper[4813]: I1206 16:40:18.764720 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-n8lrc" Dec 06 16:40:18 crc kubenswrapper[4813]: I1206 16:40:18.765121 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-n8lrc" Dec 06 16:40:18 crc kubenswrapper[4813]: I1206 16:40:18.810951 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-n8lrc" Dec 06 16:40:19 crc kubenswrapper[4813]: I1206 16:40:19.507687 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-n8lrc" Dec 06 16:40:20 crc kubenswrapper[4813]: I1206 16:40:20.023455 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n8lrc"] Dec 06 16:40:21 crc kubenswrapper[4813]: I1206 16:40:21.500066 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-n8lrc" podUID="2ad84257-a4c7-4674-9329-f862f8b1460e" containerName="registry-server" containerID="cri-o://2ba915cf88622272e091db38700585f1a8b575a946d2da70b7efcf3983a91ea5" gracePeriod=2 Dec 06 16:40:21 crc kubenswrapper[4813]: I1206 16:40:21.954688 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n8lrc" Dec 06 16:40:22 crc kubenswrapper[4813]: I1206 16:40:22.031814 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcl4k\" (UniqueName: \"kubernetes.io/projected/2ad84257-a4c7-4674-9329-f862f8b1460e-kube-api-access-zcl4k\") pod \"2ad84257-a4c7-4674-9329-f862f8b1460e\" (UID: \"2ad84257-a4c7-4674-9329-f862f8b1460e\") " Dec 06 16:40:22 crc kubenswrapper[4813]: I1206 16:40:22.032018 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ad84257-a4c7-4674-9329-f862f8b1460e-utilities\") pod \"2ad84257-a4c7-4674-9329-f862f8b1460e\" (UID: \"2ad84257-a4c7-4674-9329-f862f8b1460e\") " Dec 06 16:40:22 crc kubenswrapper[4813]: I1206 16:40:22.032180 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ad84257-a4c7-4674-9329-f862f8b1460e-catalog-content\") pod \"2ad84257-a4c7-4674-9329-f862f8b1460e\" (UID: \"2ad84257-a4c7-4674-9329-f862f8b1460e\") " Dec 06 16:40:22 crc kubenswrapper[4813]: I1206 16:40:22.032883 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ad84257-a4c7-4674-9329-f862f8b1460e-utilities" (OuterVolumeSpecName: "utilities") pod "2ad84257-a4c7-4674-9329-f862f8b1460e" (UID: "2ad84257-a4c7-4674-9329-f862f8b1460e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:40:22 crc kubenswrapper[4813]: I1206 16:40:22.037494 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ad84257-a4c7-4674-9329-f862f8b1460e-kube-api-access-zcl4k" (OuterVolumeSpecName: "kube-api-access-zcl4k") pod "2ad84257-a4c7-4674-9329-f862f8b1460e" (UID: "2ad84257-a4c7-4674-9329-f862f8b1460e"). InnerVolumeSpecName "kube-api-access-zcl4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:40:22 crc kubenswrapper[4813]: I1206 16:40:22.038945 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcl4k\" (UniqueName: \"kubernetes.io/projected/2ad84257-a4c7-4674-9329-f862f8b1460e-kube-api-access-zcl4k\") on node \"crc\" DevicePath \"\"" Dec 06 16:40:22 crc kubenswrapper[4813]: I1206 16:40:22.038979 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ad84257-a4c7-4674-9329-f862f8b1460e-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 16:40:22 crc kubenswrapper[4813]: I1206 16:40:22.094141 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ad84257-a4c7-4674-9329-f862f8b1460e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2ad84257-a4c7-4674-9329-f862f8b1460e" (UID: "2ad84257-a4c7-4674-9329-f862f8b1460e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:40:22 crc kubenswrapper[4813]: I1206 16:40:22.140365 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ad84257-a4c7-4674-9329-f862f8b1460e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 16:40:22 crc kubenswrapper[4813]: I1206 16:40:22.511005 4813 generic.go:334] "Generic (PLEG): container finished" podID="2ad84257-a4c7-4674-9329-f862f8b1460e" containerID="2ba915cf88622272e091db38700585f1a8b575a946d2da70b7efcf3983a91ea5" exitCode=0 Dec 06 16:40:22 crc kubenswrapper[4813]: I1206 16:40:22.511041 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n8lrc" event={"ID":"2ad84257-a4c7-4674-9329-f862f8b1460e","Type":"ContainerDied","Data":"2ba915cf88622272e091db38700585f1a8b575a946d2da70b7efcf3983a91ea5"} Dec 06 16:40:22 crc kubenswrapper[4813]: I1206 16:40:22.511090 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n8lrc" event={"ID":"2ad84257-a4c7-4674-9329-f862f8b1460e","Type":"ContainerDied","Data":"af4f4297a2091b8ac9befccbf99a70901b97f430e25cdf94d97c5c955f31181d"} Dec 06 16:40:22 crc kubenswrapper[4813]: I1206 16:40:22.511109 4813 scope.go:117] "RemoveContainer" containerID="2ba915cf88622272e091db38700585f1a8b575a946d2da70b7efcf3983a91ea5" Dec 06 16:40:22 crc kubenswrapper[4813]: I1206 16:40:22.511299 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n8lrc" Dec 06 16:40:22 crc kubenswrapper[4813]: I1206 16:40:22.556973 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n8lrc"] Dec 06 16:40:22 crc kubenswrapper[4813]: I1206 16:40:22.563536 4813 scope.go:117] "RemoveContainer" containerID="902280a2744bfb274515b042117c93227dd198e40c5688b88f374f488e40068d" Dec 06 16:40:22 crc kubenswrapper[4813]: I1206 16:40:22.571439 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-n8lrc"] Dec 06 16:40:22 crc kubenswrapper[4813]: I1206 16:40:22.592503 4813 scope.go:117] "RemoveContainer" containerID="df271b90b410490a6c7188e83a8a51e95be03a0d4d3af97f29a3ba8971903d80" Dec 06 16:40:22 crc kubenswrapper[4813]: I1206 16:40:22.628766 4813 scope.go:117] "RemoveContainer" containerID="2ba915cf88622272e091db38700585f1a8b575a946d2da70b7efcf3983a91ea5" Dec 06 16:40:22 crc kubenswrapper[4813]: E1206 16:40:22.629204 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ba915cf88622272e091db38700585f1a8b575a946d2da70b7efcf3983a91ea5\": container with ID starting with 2ba915cf88622272e091db38700585f1a8b575a946d2da70b7efcf3983a91ea5 not found: ID does not exist" containerID="2ba915cf88622272e091db38700585f1a8b575a946d2da70b7efcf3983a91ea5" Dec 06 16:40:22 crc kubenswrapper[4813]: I1206 16:40:22.629235 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ba915cf88622272e091db38700585f1a8b575a946d2da70b7efcf3983a91ea5"} err="failed to get container status \"2ba915cf88622272e091db38700585f1a8b575a946d2da70b7efcf3983a91ea5\": rpc error: code = NotFound desc = could not find container \"2ba915cf88622272e091db38700585f1a8b575a946d2da70b7efcf3983a91ea5\": container with ID starting with 2ba915cf88622272e091db38700585f1a8b575a946d2da70b7efcf3983a91ea5 not found: ID does not exist" Dec 06 16:40:22 crc kubenswrapper[4813]: I1206 16:40:22.629255 4813 scope.go:117] "RemoveContainer" containerID="902280a2744bfb274515b042117c93227dd198e40c5688b88f374f488e40068d" Dec 06 16:40:22 crc kubenswrapper[4813]: E1206 16:40:22.629573 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"902280a2744bfb274515b042117c93227dd198e40c5688b88f374f488e40068d\": container with ID starting with 902280a2744bfb274515b042117c93227dd198e40c5688b88f374f488e40068d not found: ID does not exist" containerID="902280a2744bfb274515b042117c93227dd198e40c5688b88f374f488e40068d" Dec 06 16:40:22 crc kubenswrapper[4813]: I1206 16:40:22.629591 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"902280a2744bfb274515b042117c93227dd198e40c5688b88f374f488e40068d"} err="failed to get container status \"902280a2744bfb274515b042117c93227dd198e40c5688b88f374f488e40068d\": rpc error: code = NotFound desc = could not find container \"902280a2744bfb274515b042117c93227dd198e40c5688b88f374f488e40068d\": container with ID starting with 902280a2744bfb274515b042117c93227dd198e40c5688b88f374f488e40068d not found: ID does not exist" Dec 06 16:40:22 crc kubenswrapper[4813]: I1206 16:40:22.629604 4813 scope.go:117] "RemoveContainer" containerID="df271b90b410490a6c7188e83a8a51e95be03a0d4d3af97f29a3ba8971903d80" Dec 06 16:40:22 crc kubenswrapper[4813]: E1206 16:40:22.629910 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df271b90b410490a6c7188e83a8a51e95be03a0d4d3af97f29a3ba8971903d80\": container with ID starting with df271b90b410490a6c7188e83a8a51e95be03a0d4d3af97f29a3ba8971903d80 not found: ID does not exist" containerID="df271b90b410490a6c7188e83a8a51e95be03a0d4d3af97f29a3ba8971903d80" Dec 06 16:40:22 crc kubenswrapper[4813]: I1206 16:40:22.629940 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df271b90b410490a6c7188e83a8a51e95be03a0d4d3af97f29a3ba8971903d80"} err="failed to get container status \"df271b90b410490a6c7188e83a8a51e95be03a0d4d3af97f29a3ba8971903d80\": rpc error: code = NotFound desc = could not find container \"df271b90b410490a6c7188e83a8a51e95be03a0d4d3af97f29a3ba8971903d80\": container with ID starting with df271b90b410490a6c7188e83a8a51e95be03a0d4d3af97f29a3ba8971903d80 not found: ID does not exist" Dec 06 16:40:24 crc kubenswrapper[4813]: I1206 16:40:24.496745 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ad84257-a4c7-4674-9329-f862f8b1460e" path="/var/lib/kubelet/pods/2ad84257-a4c7-4674-9329-f862f8b1460e/volumes" Dec 06 16:40:25 crc kubenswrapper[4813]: I1206 16:40:25.487200 4813 scope.go:117] "RemoveContainer" containerID="8441b376fae9405da00e04a7e75ce187d0f25185f6fc0f6fc4224d7561e835bd" Dec 06 16:40:26 crc kubenswrapper[4813]: I1206 16:40:26.551832 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerStarted","Data":"89167fab7a057df4dbc2e1e5aa930f5ed89747de01e89ff0bb679d87ca1d7a24"} Dec 06 16:40:38 crc kubenswrapper[4813]: I1206 16:40:38.670375 4813 generic.go:334] "Generic (PLEG): container finished" podID="1e7de122-b529-4ee9-ac39-a31a22fc19e9" containerID="e8eeef2364d649555a536ddd16a7b6a11037a5c4688dae5922f1f23c4a727276" exitCode=0 Dec 06 16:40:38 crc kubenswrapper[4813]: I1206 16:40:38.670497 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jlgf8/crc-debug-jnnkw" event={"ID":"1e7de122-b529-4ee9-ac39-a31a22fc19e9","Type":"ContainerDied","Data":"e8eeef2364d649555a536ddd16a7b6a11037a5c4688dae5922f1f23c4a727276"} Dec 06 16:40:39 crc kubenswrapper[4813]: I1206 16:40:39.809602 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jlgf8/crc-debug-jnnkw" Dec 06 16:40:39 crc kubenswrapper[4813]: I1206 16:40:39.863745 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jlgf8/crc-debug-jnnkw"] Dec 06 16:40:39 crc kubenswrapper[4813]: I1206 16:40:39.874852 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jlgf8/crc-debug-jnnkw"] Dec 06 16:40:39 crc kubenswrapper[4813]: I1206 16:40:39.905186 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1e7de122-b529-4ee9-ac39-a31a22fc19e9-host\") pod \"1e7de122-b529-4ee9-ac39-a31a22fc19e9\" (UID: \"1e7de122-b529-4ee9-ac39-a31a22fc19e9\") " Dec 06 16:40:39 crc kubenswrapper[4813]: I1206 16:40:39.905508 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pwc4\" (UniqueName: \"kubernetes.io/projected/1e7de122-b529-4ee9-ac39-a31a22fc19e9-kube-api-access-2pwc4\") pod \"1e7de122-b529-4ee9-ac39-a31a22fc19e9\" (UID: \"1e7de122-b529-4ee9-ac39-a31a22fc19e9\") " Dec 06 16:40:39 crc kubenswrapper[4813]: I1206 16:40:39.905390 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1e7de122-b529-4ee9-ac39-a31a22fc19e9-host" (OuterVolumeSpecName: "host") pod "1e7de122-b529-4ee9-ac39-a31a22fc19e9" (UID: "1e7de122-b529-4ee9-ac39-a31a22fc19e9"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 16:40:39 crc kubenswrapper[4813]: I1206 16:40:39.906159 4813 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1e7de122-b529-4ee9-ac39-a31a22fc19e9-host\") on node \"crc\" DevicePath \"\"" Dec 06 16:40:39 crc kubenswrapper[4813]: I1206 16:40:39.911981 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e7de122-b529-4ee9-ac39-a31a22fc19e9-kube-api-access-2pwc4" (OuterVolumeSpecName: "kube-api-access-2pwc4") pod "1e7de122-b529-4ee9-ac39-a31a22fc19e9" (UID: "1e7de122-b529-4ee9-ac39-a31a22fc19e9"). InnerVolumeSpecName "kube-api-access-2pwc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:40:40 crc kubenswrapper[4813]: I1206 16:40:40.008366 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pwc4\" (UniqueName: \"kubernetes.io/projected/1e7de122-b529-4ee9-ac39-a31a22fc19e9-kube-api-access-2pwc4\") on node \"crc\" DevicePath \"\"" Dec 06 16:40:40 crc kubenswrapper[4813]: I1206 16:40:40.508089 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e7de122-b529-4ee9-ac39-a31a22fc19e9" path="/var/lib/kubelet/pods/1e7de122-b529-4ee9-ac39-a31a22fc19e9/volumes" Dec 06 16:40:40 crc kubenswrapper[4813]: I1206 16:40:40.694732 4813 scope.go:117] "RemoveContainer" containerID="e8eeef2364d649555a536ddd16a7b6a11037a5c4688dae5922f1f23c4a727276" Dec 06 16:40:40 crc kubenswrapper[4813]: I1206 16:40:40.694953 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jlgf8/crc-debug-jnnkw" Dec 06 16:40:41 crc kubenswrapper[4813]: I1206 16:40:41.050072 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jlgf8/crc-debug-ng9ks"] Dec 06 16:40:41 crc kubenswrapper[4813]: E1206 16:40:41.052488 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ad84257-a4c7-4674-9329-f862f8b1460e" containerName="extract-content" Dec 06 16:40:41 crc kubenswrapper[4813]: I1206 16:40:41.052509 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ad84257-a4c7-4674-9329-f862f8b1460e" containerName="extract-content" Dec 06 16:40:41 crc kubenswrapper[4813]: E1206 16:40:41.052525 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e7de122-b529-4ee9-ac39-a31a22fc19e9" containerName="container-00" Dec 06 16:40:41 crc kubenswrapper[4813]: I1206 16:40:41.052531 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e7de122-b529-4ee9-ac39-a31a22fc19e9" containerName="container-00" Dec 06 16:40:41 crc kubenswrapper[4813]: E1206 16:40:41.052540 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ad84257-a4c7-4674-9329-f862f8b1460e" containerName="extract-utilities" Dec 06 16:40:41 crc kubenswrapper[4813]: I1206 16:40:41.052547 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ad84257-a4c7-4674-9329-f862f8b1460e" containerName="extract-utilities" Dec 06 16:40:41 crc kubenswrapper[4813]: E1206 16:40:41.052563 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ad84257-a4c7-4674-9329-f862f8b1460e" containerName="registry-server" Dec 06 16:40:41 crc kubenswrapper[4813]: I1206 16:40:41.052568 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ad84257-a4c7-4674-9329-f862f8b1460e" containerName="registry-server" Dec 06 16:40:41 crc kubenswrapper[4813]: I1206 16:40:41.052730 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e7de122-b529-4ee9-ac39-a31a22fc19e9" containerName="container-00" Dec 06 16:40:41 crc kubenswrapper[4813]: I1206 16:40:41.052747 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ad84257-a4c7-4674-9329-f862f8b1460e" containerName="registry-server" Dec 06 16:40:41 crc kubenswrapper[4813]: I1206 16:40:41.053328 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jlgf8/crc-debug-ng9ks" Dec 06 16:40:41 crc kubenswrapper[4813]: I1206 16:40:41.055818 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-jlgf8"/"default-dockercfg-s4b7p" Dec 06 16:40:41 crc kubenswrapper[4813]: I1206 16:40:41.057524 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/429ccdb6-935f-4d75-b1b1-cd12e65cbf20-host\") pod \"crc-debug-ng9ks\" (UID: \"429ccdb6-935f-4d75-b1b1-cd12e65cbf20\") " pod="openshift-must-gather-jlgf8/crc-debug-ng9ks" Dec 06 16:40:41 crc kubenswrapper[4813]: I1206 16:40:41.057622 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrjtw\" (UniqueName: \"kubernetes.io/projected/429ccdb6-935f-4d75-b1b1-cd12e65cbf20-kube-api-access-mrjtw\") pod \"crc-debug-ng9ks\" (UID: \"429ccdb6-935f-4d75-b1b1-cd12e65cbf20\") " pod="openshift-must-gather-jlgf8/crc-debug-ng9ks" Dec 06 16:40:41 crc kubenswrapper[4813]: I1206 16:40:41.160232 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/429ccdb6-935f-4d75-b1b1-cd12e65cbf20-host\") pod \"crc-debug-ng9ks\" (UID: \"429ccdb6-935f-4d75-b1b1-cd12e65cbf20\") " pod="openshift-must-gather-jlgf8/crc-debug-ng9ks" Dec 06 16:40:41 crc kubenswrapper[4813]: I1206 16:40:41.160458 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/429ccdb6-935f-4d75-b1b1-cd12e65cbf20-host\") pod \"crc-debug-ng9ks\" (UID: \"429ccdb6-935f-4d75-b1b1-cd12e65cbf20\") " pod="openshift-must-gather-jlgf8/crc-debug-ng9ks" Dec 06 16:40:41 crc kubenswrapper[4813]: I1206 16:40:41.160536 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrjtw\" (UniqueName: \"kubernetes.io/projected/429ccdb6-935f-4d75-b1b1-cd12e65cbf20-kube-api-access-mrjtw\") pod \"crc-debug-ng9ks\" (UID: \"429ccdb6-935f-4d75-b1b1-cd12e65cbf20\") " pod="openshift-must-gather-jlgf8/crc-debug-ng9ks" Dec 06 16:40:41 crc kubenswrapper[4813]: I1206 16:40:41.196125 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrjtw\" (UniqueName: \"kubernetes.io/projected/429ccdb6-935f-4d75-b1b1-cd12e65cbf20-kube-api-access-mrjtw\") pod \"crc-debug-ng9ks\" (UID: \"429ccdb6-935f-4d75-b1b1-cd12e65cbf20\") " pod="openshift-must-gather-jlgf8/crc-debug-ng9ks" Dec 06 16:40:41 crc kubenswrapper[4813]: I1206 16:40:41.374740 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jlgf8/crc-debug-ng9ks" Dec 06 16:40:41 crc kubenswrapper[4813]: W1206 16:40:41.419633 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod429ccdb6_935f_4d75_b1b1_cd12e65cbf20.slice/crio-4ebef5b8b0cc0cd89c6ba11e1b0f240dbd1b4b83c461ebd85846f48cdb87a1e9 WatchSource:0}: Error finding container 4ebef5b8b0cc0cd89c6ba11e1b0f240dbd1b4b83c461ebd85846f48cdb87a1e9: Status 404 returned error can't find the container with id 4ebef5b8b0cc0cd89c6ba11e1b0f240dbd1b4b83c461ebd85846f48cdb87a1e9 Dec 06 16:40:41 crc kubenswrapper[4813]: I1206 16:40:41.729489 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jlgf8/crc-debug-ng9ks" event={"ID":"429ccdb6-935f-4d75-b1b1-cd12e65cbf20","Type":"ContainerStarted","Data":"611934eee86a60d5c24976ab0fb4239a265bae3a3c2cdf13ddeca19bd33000bd"} Dec 06 16:40:41 crc kubenswrapper[4813]: I1206 16:40:41.729536 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jlgf8/crc-debug-ng9ks" event={"ID":"429ccdb6-935f-4d75-b1b1-cd12e65cbf20","Type":"ContainerStarted","Data":"4ebef5b8b0cc0cd89c6ba11e1b0f240dbd1b4b83c461ebd85846f48cdb87a1e9"} Dec 06 16:40:41 crc kubenswrapper[4813]: I1206 16:40:41.750273 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-jlgf8/crc-debug-ng9ks" podStartSLOduration=0.750237708 podStartE2EDuration="750.237708ms" podCreationTimestamp="2025-12-06 16:40:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:40:41.748523983 +0000 UTC m=+3281.639403559" watchObservedRunningTime="2025-12-06 16:40:41.750237708 +0000 UTC m=+3281.641117294" Dec 06 16:40:42 crc kubenswrapper[4813]: I1206 16:40:42.746774 4813 generic.go:334] "Generic (PLEG): container finished" podID="429ccdb6-935f-4d75-b1b1-cd12e65cbf20" containerID="611934eee86a60d5c24976ab0fb4239a265bae3a3c2cdf13ddeca19bd33000bd" exitCode=0 Dec 06 16:40:42 crc kubenswrapper[4813]: I1206 16:40:42.746829 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jlgf8/crc-debug-ng9ks" event={"ID":"429ccdb6-935f-4d75-b1b1-cd12e65cbf20","Type":"ContainerDied","Data":"611934eee86a60d5c24976ab0fb4239a265bae3a3c2cdf13ddeca19bd33000bd"} Dec 06 16:40:43 crc kubenswrapper[4813]: I1206 16:40:43.853526 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jlgf8/crc-debug-ng9ks" Dec 06 16:40:43 crc kubenswrapper[4813]: I1206 16:40:43.892025 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jlgf8/crc-debug-ng9ks"] Dec 06 16:40:43 crc kubenswrapper[4813]: I1206 16:40:43.900957 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jlgf8/crc-debug-ng9ks"] Dec 06 16:40:43 crc kubenswrapper[4813]: I1206 16:40:43.914889 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrjtw\" (UniqueName: \"kubernetes.io/projected/429ccdb6-935f-4d75-b1b1-cd12e65cbf20-kube-api-access-mrjtw\") pod \"429ccdb6-935f-4d75-b1b1-cd12e65cbf20\" (UID: \"429ccdb6-935f-4d75-b1b1-cd12e65cbf20\") " Dec 06 16:40:43 crc kubenswrapper[4813]: I1206 16:40:43.915058 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/429ccdb6-935f-4d75-b1b1-cd12e65cbf20-host\") pod \"429ccdb6-935f-4d75-b1b1-cd12e65cbf20\" (UID: \"429ccdb6-935f-4d75-b1b1-cd12e65cbf20\") " Dec 06 16:40:43 crc kubenswrapper[4813]: I1206 16:40:43.915240 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/429ccdb6-935f-4d75-b1b1-cd12e65cbf20-host" (OuterVolumeSpecName: "host") pod "429ccdb6-935f-4d75-b1b1-cd12e65cbf20" (UID: "429ccdb6-935f-4d75-b1b1-cd12e65cbf20"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 16:40:43 crc kubenswrapper[4813]: I1206 16:40:43.915602 4813 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/429ccdb6-935f-4d75-b1b1-cd12e65cbf20-host\") on node \"crc\" DevicePath \"\"" Dec 06 16:40:43 crc kubenswrapper[4813]: I1206 16:40:43.923834 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/429ccdb6-935f-4d75-b1b1-cd12e65cbf20-kube-api-access-mrjtw" (OuterVolumeSpecName: "kube-api-access-mrjtw") pod "429ccdb6-935f-4d75-b1b1-cd12e65cbf20" (UID: "429ccdb6-935f-4d75-b1b1-cd12e65cbf20"). InnerVolumeSpecName "kube-api-access-mrjtw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:40:44 crc kubenswrapper[4813]: I1206 16:40:44.016711 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrjtw\" (UniqueName: \"kubernetes.io/projected/429ccdb6-935f-4d75-b1b1-cd12e65cbf20-kube-api-access-mrjtw\") on node \"crc\" DevicePath \"\"" Dec 06 16:40:44 crc kubenswrapper[4813]: I1206 16:40:44.507010 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="429ccdb6-935f-4d75-b1b1-cd12e65cbf20" path="/var/lib/kubelet/pods/429ccdb6-935f-4d75-b1b1-cd12e65cbf20/volumes" Dec 06 16:40:44 crc kubenswrapper[4813]: I1206 16:40:44.773410 4813 scope.go:117] "RemoveContainer" containerID="611934eee86a60d5c24976ab0fb4239a265bae3a3c2cdf13ddeca19bd33000bd" Dec 06 16:40:44 crc kubenswrapper[4813]: I1206 16:40:44.773582 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jlgf8/crc-debug-ng9ks" Dec 06 16:40:45 crc kubenswrapper[4813]: I1206 16:40:45.136312 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jlgf8/crc-debug-jgcbq"] Dec 06 16:40:45 crc kubenswrapper[4813]: E1206 16:40:45.136749 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="429ccdb6-935f-4d75-b1b1-cd12e65cbf20" containerName="container-00" Dec 06 16:40:45 crc kubenswrapper[4813]: I1206 16:40:45.136763 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="429ccdb6-935f-4d75-b1b1-cd12e65cbf20" containerName="container-00" Dec 06 16:40:45 crc kubenswrapper[4813]: I1206 16:40:45.136991 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="429ccdb6-935f-4d75-b1b1-cd12e65cbf20" containerName="container-00" Dec 06 16:40:45 crc kubenswrapper[4813]: I1206 16:40:45.137786 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jlgf8/crc-debug-jgcbq" Dec 06 16:40:45 crc kubenswrapper[4813]: I1206 16:40:45.140194 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-jlgf8"/"default-dockercfg-s4b7p" Dec 06 16:40:45 crc kubenswrapper[4813]: I1206 16:40:45.341316 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crmp2\" (UniqueName: \"kubernetes.io/projected/0a7a9120-6a77-47fd-abfe-4f093e1baf1e-kube-api-access-crmp2\") pod \"crc-debug-jgcbq\" (UID: \"0a7a9120-6a77-47fd-abfe-4f093e1baf1e\") " pod="openshift-must-gather-jlgf8/crc-debug-jgcbq" Dec 06 16:40:45 crc kubenswrapper[4813]: I1206 16:40:45.341369 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0a7a9120-6a77-47fd-abfe-4f093e1baf1e-host\") pod \"crc-debug-jgcbq\" (UID: \"0a7a9120-6a77-47fd-abfe-4f093e1baf1e\") " pod="openshift-must-gather-jlgf8/crc-debug-jgcbq" Dec 06 16:40:45 crc kubenswrapper[4813]: I1206 16:40:45.443081 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crmp2\" (UniqueName: \"kubernetes.io/projected/0a7a9120-6a77-47fd-abfe-4f093e1baf1e-kube-api-access-crmp2\") pod \"crc-debug-jgcbq\" (UID: \"0a7a9120-6a77-47fd-abfe-4f093e1baf1e\") " pod="openshift-must-gather-jlgf8/crc-debug-jgcbq" Dec 06 16:40:45 crc kubenswrapper[4813]: I1206 16:40:45.443156 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0a7a9120-6a77-47fd-abfe-4f093e1baf1e-host\") pod \"crc-debug-jgcbq\" (UID: \"0a7a9120-6a77-47fd-abfe-4f093e1baf1e\") " pod="openshift-must-gather-jlgf8/crc-debug-jgcbq" Dec 06 16:40:45 crc kubenswrapper[4813]: I1206 16:40:45.443407 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0a7a9120-6a77-47fd-abfe-4f093e1baf1e-host\") pod \"crc-debug-jgcbq\" (UID: \"0a7a9120-6a77-47fd-abfe-4f093e1baf1e\") " pod="openshift-must-gather-jlgf8/crc-debug-jgcbq" Dec 06 16:40:45 crc kubenswrapper[4813]: I1206 16:40:45.467593 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crmp2\" (UniqueName: \"kubernetes.io/projected/0a7a9120-6a77-47fd-abfe-4f093e1baf1e-kube-api-access-crmp2\") pod \"crc-debug-jgcbq\" (UID: \"0a7a9120-6a77-47fd-abfe-4f093e1baf1e\") " pod="openshift-must-gather-jlgf8/crc-debug-jgcbq" Dec 06 16:40:45 crc kubenswrapper[4813]: I1206 16:40:45.756954 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jlgf8/crc-debug-jgcbq" Dec 06 16:40:46 crc kubenswrapper[4813]: I1206 16:40:46.817893 4813 generic.go:334] "Generic (PLEG): container finished" podID="0a7a9120-6a77-47fd-abfe-4f093e1baf1e" containerID="53da53da50a60415767f89c445891954582fd2dd2b2418ff79efc798eeb9dee6" exitCode=0 Dec 06 16:40:46 crc kubenswrapper[4813]: I1206 16:40:46.818194 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jlgf8/crc-debug-jgcbq" event={"ID":"0a7a9120-6a77-47fd-abfe-4f093e1baf1e","Type":"ContainerDied","Data":"53da53da50a60415767f89c445891954582fd2dd2b2418ff79efc798eeb9dee6"} Dec 06 16:40:46 crc kubenswrapper[4813]: I1206 16:40:46.818239 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jlgf8/crc-debug-jgcbq" event={"ID":"0a7a9120-6a77-47fd-abfe-4f093e1baf1e","Type":"ContainerStarted","Data":"9cbbf1f209661eec7344e73cea494b0c8c2546f88274ed5e7ab74cc04d72314e"} Dec 06 16:40:46 crc kubenswrapper[4813]: I1206 16:40:46.867823 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jlgf8/crc-debug-jgcbq"] Dec 06 16:40:46 crc kubenswrapper[4813]: I1206 16:40:46.874912 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jlgf8/crc-debug-jgcbq"] Dec 06 16:40:47 crc kubenswrapper[4813]: I1206 16:40:47.959605 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jlgf8/crc-debug-jgcbq" Dec 06 16:40:48 crc kubenswrapper[4813]: I1206 16:40:48.100072 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crmp2\" (UniqueName: \"kubernetes.io/projected/0a7a9120-6a77-47fd-abfe-4f093e1baf1e-kube-api-access-crmp2\") pod \"0a7a9120-6a77-47fd-abfe-4f093e1baf1e\" (UID: \"0a7a9120-6a77-47fd-abfe-4f093e1baf1e\") " Dec 06 16:40:48 crc kubenswrapper[4813]: I1206 16:40:48.100598 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0a7a9120-6a77-47fd-abfe-4f093e1baf1e-host\") pod \"0a7a9120-6a77-47fd-abfe-4f093e1baf1e\" (UID: \"0a7a9120-6a77-47fd-abfe-4f093e1baf1e\") " Dec 06 16:40:48 crc kubenswrapper[4813]: I1206 16:40:48.100708 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0a7a9120-6a77-47fd-abfe-4f093e1baf1e-host" (OuterVolumeSpecName: "host") pod "0a7a9120-6a77-47fd-abfe-4f093e1baf1e" (UID: "0a7a9120-6a77-47fd-abfe-4f093e1baf1e"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 16:40:48 crc kubenswrapper[4813]: I1206 16:40:48.101536 4813 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0a7a9120-6a77-47fd-abfe-4f093e1baf1e-host\") on node \"crc\" DevicePath \"\"" Dec 06 16:40:48 crc kubenswrapper[4813]: I1206 16:40:48.105617 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a7a9120-6a77-47fd-abfe-4f093e1baf1e-kube-api-access-crmp2" (OuterVolumeSpecName: "kube-api-access-crmp2") pod "0a7a9120-6a77-47fd-abfe-4f093e1baf1e" (UID: "0a7a9120-6a77-47fd-abfe-4f093e1baf1e"). InnerVolumeSpecName "kube-api-access-crmp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:40:48 crc kubenswrapper[4813]: I1206 16:40:48.202638 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crmp2\" (UniqueName: \"kubernetes.io/projected/0a7a9120-6a77-47fd-abfe-4f093e1baf1e-kube-api-access-crmp2\") on node \"crc\" DevicePath \"\"" Dec 06 16:40:48 crc kubenswrapper[4813]: I1206 16:40:48.498214 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a7a9120-6a77-47fd-abfe-4f093e1baf1e" path="/var/lib/kubelet/pods/0a7a9120-6a77-47fd-abfe-4f093e1baf1e/volumes" Dec 06 16:40:48 crc kubenswrapper[4813]: I1206 16:40:48.836874 4813 scope.go:117] "RemoveContainer" containerID="53da53da50a60415767f89c445891954582fd2dd2b2418ff79efc798eeb9dee6" Dec 06 16:40:48 crc kubenswrapper[4813]: I1206 16:40:48.836919 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jlgf8/crc-debug-jgcbq" Dec 06 16:41:05 crc kubenswrapper[4813]: I1206 16:41:05.697381 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-58d4d548d-6bl8w_f883aaa3-f010-4c52-bc12-1591fb0ddb4d/barbican-api/0.log" Dec 06 16:41:06 crc kubenswrapper[4813]: I1206 16:41:06.073877 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-64449f75dd-k89r7_322df0dd-86e6-466b-8afa-e4e52e889827/barbican-keystone-listener/0.log" Dec 06 16:41:06 crc kubenswrapper[4813]: I1206 16:41:06.329051 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-64449f75dd-k89r7_322df0dd-86e6-466b-8afa-e4e52e889827/barbican-keystone-listener-log/0.log" Dec 06 16:41:06 crc kubenswrapper[4813]: I1206 16:41:06.363205 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-58d4d548d-6bl8w_f883aaa3-f010-4c52-bc12-1591fb0ddb4d/barbican-api-log/0.log" Dec 06 16:41:06 crc kubenswrapper[4813]: I1206 16:41:06.408851 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-ffc86b7c-26pxs_cf88c989-e78b-456e-ba32-19d8af98a1d0/barbican-worker/0.log" Dec 06 16:41:06 crc kubenswrapper[4813]: I1206 16:41:06.653659 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-ffc86b7c-26pxs_cf88c989-e78b-456e-ba32-19d8af98a1d0/barbican-worker-log/0.log" Dec 06 16:41:06 crc kubenswrapper[4813]: I1206 16:41:06.672225 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-qr892_6306b1cf-bf21-4cd0-a7c8-c83a99863266/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:41:06 crc kubenswrapper[4813]: I1206 16:41:06.810195 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_66613ee6-b0b7-48fc-adb6-86bab4940550/ceilometer-central-agent/0.log" Dec 06 16:41:06 crc kubenswrapper[4813]: I1206 16:41:06.899672 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_66613ee6-b0b7-48fc-adb6-86bab4940550/ceilometer-notification-agent/0.log" Dec 06 16:41:06 crc kubenswrapper[4813]: I1206 16:41:06.954107 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_66613ee6-b0b7-48fc-adb6-86bab4940550/proxy-httpd/0.log" Dec 06 16:41:07 crc kubenswrapper[4813]: I1206 16:41:07.003104 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_66613ee6-b0b7-48fc-adb6-86bab4940550/sg-core/0.log" Dec 06 16:41:07 crc kubenswrapper[4813]: I1206 16:41:07.191691 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d/cinder-api-log/0.log" Dec 06 16:41:07 crc kubenswrapper[4813]: I1206 16:41:07.198636 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d/cinder-api/0.log" Dec 06 16:41:07 crc kubenswrapper[4813]: I1206 16:41:07.617398 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_7d707d75-8ac3-4256-9e9e-018e5caef917/cinder-scheduler/0.log" Dec 06 16:41:07 crc kubenswrapper[4813]: I1206 16:41:07.703509 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-48kg5_11276539-afd1-4649-97eb-867dfcc76819/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:41:07 crc kubenswrapper[4813]: I1206 16:41:07.719723 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_7d707d75-8ac3-4256-9e9e-018e5caef917/probe/0.log" Dec 06 16:41:08 crc kubenswrapper[4813]: I1206 16:41:08.013864 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl_6f525ab6-f23e-4cc9-a843-7a601896c864/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:41:08 crc kubenswrapper[4813]: I1206 16:41:08.037539 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6b6dc74c5-4w85d_5399f130-d1b7-48ba-a32b-6da83416719e/init/0.log" Dec 06 16:41:08 crc kubenswrapper[4813]: I1206 16:41:08.350832 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6b6dc74c5-4w85d_5399f130-d1b7-48ba-a32b-6da83416719e/dnsmasq-dns/0.log" Dec 06 16:41:08 crc kubenswrapper[4813]: I1206 16:41:08.390954 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h_c2bb7ef8-9c22-412d-83a9-686036ecb3f8/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:41:08 crc kubenswrapper[4813]: I1206 16:41:08.416254 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6b6dc74c5-4w85d_5399f130-d1b7-48ba-a32b-6da83416719e/init/0.log" Dec 06 16:41:08 crc kubenswrapper[4813]: I1206 16:41:08.679729 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_ee3e963f-6982-4cd8-87d2-4eceaec10be3/glance-httpd/0.log" Dec 06 16:41:08 crc kubenswrapper[4813]: I1206 16:41:08.703925 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_ee3e963f-6982-4cd8-87d2-4eceaec10be3/glance-log/0.log" Dec 06 16:41:08 crc kubenswrapper[4813]: I1206 16:41:08.965480 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_ec17345c-3d8e-4981-945c-173fb39aab99/glance-httpd/0.log" Dec 06 16:41:09 crc kubenswrapper[4813]: I1206 16:41:09.031347 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_ec17345c-3d8e-4981-945c-173fb39aab99/glance-log/0.log" Dec 06 16:41:09 crc kubenswrapper[4813]: I1206 16:41:09.096381 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7fd6f7946b-kmm96_28238998-5d44-4ae9-8d1f-93d56ff18152/horizon/0.log" Dec 06 16:41:09 crc kubenswrapper[4813]: I1206 16:41:09.442149 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7fd6f7946b-kmm96_28238998-5d44-4ae9-8d1f-93d56ff18152/horizon-log/0.log" Dec 06 16:41:09 crc kubenswrapper[4813]: I1206 16:41:09.448281 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg_8c3bde1a-68be-4aaf-9c46-f5ea14578783/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:41:09 crc kubenswrapper[4813]: I1206 16:41:09.622360 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-nbw6k_c8332abd-e13b-4587-831b-1330e18de573/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:41:09 crc kubenswrapper[4813]: I1206 16:41:09.834864 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-c4f8d8d9b-p5pqq_c6d636bb-c76a-4895-8af6-6801f45e4e5b/keystone-api/0.log" Dec 06 16:41:09 crc kubenswrapper[4813]: I1206 16:41:09.878419 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_8d3a2b2d-c016-4948-83c5-f8fcae8678d8/kube-state-metrics/0.log" Dec 06 16:41:10 crc kubenswrapper[4813]: I1206 16:41:10.081292 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j_bfa11078-d153-43a2-abc4-b70f491dc4a6/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:41:10 crc kubenswrapper[4813]: I1206 16:41:10.472754 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-759b8b45fc-psf25_afd03411-dfea-469b-804c-7cbe8a964d54/neutron-api/0.log" Dec 06 16:41:10 crc kubenswrapper[4813]: I1206 16:41:10.492021 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk_a11c7f97-21d3-48b1-9640-529df9d946b8/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:41:10 crc kubenswrapper[4813]: I1206 16:41:10.598025 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-759b8b45fc-psf25_afd03411-dfea-469b-804c-7cbe8a964d54/neutron-httpd/0.log" Dec 06 16:41:11 crc kubenswrapper[4813]: I1206 16:41:11.029898 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_e54ec1f1-806c-475d-843b-53a55fea0fab/nova-api-log/0.log" Dec 06 16:41:11 crc kubenswrapper[4813]: I1206 16:41:11.122373 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_e54ec1f1-806c-475d-843b-53a55fea0fab/nova-api-api/0.log" Dec 06 16:41:11 crc kubenswrapper[4813]: I1206 16:41:11.236658 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_db38328d-a236-4b36-bb91-9a9b818d39f1/nova-cell0-conductor-conductor/0.log" Dec 06 16:41:11 crc kubenswrapper[4813]: I1206 16:41:11.399276 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_0a45c584-458c-4f14-b7f5-50711d94207c/nova-cell1-conductor-conductor/0.log" Dec 06 16:41:11 crc kubenswrapper[4813]: I1206 16:41:11.556326 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_e9d64e47-211f-4eab-84e9-0caabeff895b/nova-cell1-novncproxy-novncproxy/0.log" Dec 06 16:41:11 crc kubenswrapper[4813]: I1206 16:41:11.737772 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-mfh5s_7ba82e65-8192-4943-8a2a-863d041ad231/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:41:11 crc kubenswrapper[4813]: I1206 16:41:11.957289 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_1731c7e7-8ded-4a61-8b8a-d94cf8b4a860/nova-metadata-log/0.log" Dec 06 16:41:12 crc kubenswrapper[4813]: I1206 16:41:12.242059 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_d53150b5-8980-47ff-bc3a-6b75ac75fa14/nova-scheduler-scheduler/0.log" Dec 06 16:41:12 crc kubenswrapper[4813]: I1206 16:41:12.374651 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_bf7377f8-ef23-484a-b66b-90db1cd5b7cd/mysql-bootstrap/0.log" Dec 06 16:41:12 crc kubenswrapper[4813]: I1206 16:41:12.649321 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_bf7377f8-ef23-484a-b66b-90db1cd5b7cd/mysql-bootstrap/0.log" Dec 06 16:41:12 crc kubenswrapper[4813]: I1206 16:41:12.696547 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_bf7377f8-ef23-484a-b66b-90db1cd5b7cd/galera/0.log" Dec 06 16:41:12 crc kubenswrapper[4813]: I1206 16:41:12.827628 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_75d94fff-d912-4493-b9b9-c458138a7ccf/mysql-bootstrap/0.log" Dec 06 16:41:12 crc kubenswrapper[4813]: I1206 16:41:12.939221 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_1731c7e7-8ded-4a61-8b8a-d94cf8b4a860/nova-metadata-metadata/0.log" Dec 06 16:41:13 crc kubenswrapper[4813]: I1206 16:41:13.103729 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_75d94fff-d912-4493-b9b9-c458138a7ccf/mysql-bootstrap/0.log" Dec 06 16:41:13 crc kubenswrapper[4813]: I1206 16:41:13.478525 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_75d94fff-d912-4493-b9b9-c458138a7ccf/galera/0.log" Dec 06 16:41:13 crc kubenswrapper[4813]: I1206 16:41:13.537483 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_0447affb-1196-4c65-8e3d-936864d4fdba/openstackclient/0.log" Dec 06 16:41:13 crc kubenswrapper[4813]: I1206 16:41:13.873082 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-9bcrr_da431468-e795-477e-97cd-b1796a76a117/openstack-network-exporter/0.log" Dec 06 16:41:13 crc kubenswrapper[4813]: I1206 16:41:13.895554 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-bg6cb_ba6825c1-2137-46d6-a695-6bdd53a13ec4/ovn-controller/0.log" Dec 06 16:41:14 crc kubenswrapper[4813]: I1206 16:41:14.119248 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mk4nr_6b7cce06-d3db-47cd-a542-a77bfdc42f82/ovsdb-server-init/0.log" Dec 06 16:41:14 crc kubenswrapper[4813]: I1206 16:41:14.581438 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mk4nr_6b7cce06-d3db-47cd-a542-a77bfdc42f82/ovsdb-server/0.log" Dec 06 16:41:14 crc kubenswrapper[4813]: I1206 16:41:14.646922 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mk4nr_6b7cce06-d3db-47cd-a542-a77bfdc42f82/ovsdb-server-init/0.log" Dec 06 16:41:14 crc kubenswrapper[4813]: I1206 16:41:14.777856 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mk4nr_6b7cce06-d3db-47cd-a542-a77bfdc42f82/ovs-vswitchd/0.log" Dec 06 16:41:15 crc kubenswrapper[4813]: I1206 16:41:15.012218 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-6gtnd_ef88d823-7e10-4eee-a6b1-66c680355906/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:41:15 crc kubenswrapper[4813]: I1206 16:41:15.035774 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_9614929b-2a14-4d49-8009-8f627fb000fd/openstack-network-exporter/0.log" Dec 06 16:41:15 crc kubenswrapper[4813]: I1206 16:41:15.133752 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_9614929b-2a14-4d49-8009-8f627fb000fd/ovn-northd/0.log" Dec 06 16:41:15 crc kubenswrapper[4813]: I1206 16:41:15.345382 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c0a272bd-6444-4c90-acb0-fb653bf3b180/openstack-network-exporter/0.log" Dec 06 16:41:15 crc kubenswrapper[4813]: I1206 16:41:15.602693 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c0a272bd-6444-4c90-acb0-fb653bf3b180/ovsdbserver-nb/0.log" Dec 06 16:41:15 crc kubenswrapper[4813]: I1206 16:41:15.678045 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420/openstack-network-exporter/0.log" Dec 06 16:41:15 crc kubenswrapper[4813]: I1206 16:41:15.763417 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420/ovsdbserver-sb/0.log" Dec 06 16:41:15 crc kubenswrapper[4813]: I1206 16:41:15.992834 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-98dddfffb-l698f_e913dfcd-dfe3-4c2c-b72a-8393295c95fc/placement-api/0.log" Dec 06 16:41:16 crc kubenswrapper[4813]: I1206 16:41:16.211443 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-98dddfffb-l698f_e913dfcd-dfe3-4c2c-b72a-8393295c95fc/placement-log/0.log" Dec 06 16:41:16 crc kubenswrapper[4813]: I1206 16:41:16.684460 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_342df504-7a38-4c50-bb3d-8c7b28521e64/setup-container/0.log" Dec 06 16:41:16 crc kubenswrapper[4813]: I1206 16:41:16.970243 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_342df504-7a38-4c50-bb3d-8c7b28521e64/setup-container/0.log" Dec 06 16:41:17 crc kubenswrapper[4813]: I1206 16:41:17.031721 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_5ef242c1-0127-4963-9563-4943a584398c/setup-container/0.log" Dec 06 16:41:17 crc kubenswrapper[4813]: I1206 16:41:17.101989 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_342df504-7a38-4c50-bb3d-8c7b28521e64/rabbitmq/0.log" Dec 06 16:41:17 crc kubenswrapper[4813]: I1206 16:41:17.279497 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_5ef242c1-0127-4963-9563-4943a584398c/setup-container/0.log" Dec 06 16:41:17 crc kubenswrapper[4813]: I1206 16:41:17.298390 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_5ef242c1-0127-4963-9563-4943a584398c/rabbitmq/0.log" Dec 06 16:41:17 crc kubenswrapper[4813]: I1206 16:41:17.437694 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5_0ddaea40-3167-4f8b-8de0-756048946c8c/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:41:17 crc kubenswrapper[4813]: I1206 16:41:17.643060 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-n6p6l_aa40ab8b-bc66-4e51-a126-ec15a6250486/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:41:17 crc kubenswrapper[4813]: I1206 16:41:17.968424 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2_ef5432f4-56a6-4518-a6b1-00ac9b29293f/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:41:18 crc kubenswrapper[4813]: I1206 16:41:18.063157 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-h7rjq_97035b28-6546-4c6b-b273-085f7ca5bb4f/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:41:18 crc kubenswrapper[4813]: I1206 16:41:18.371895 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-nj5n7_41e561aa-3484-4946-a8d9-174ea9a22a45/ssh-known-hosts-edpm-deployment/0.log" Dec 06 16:41:18 crc kubenswrapper[4813]: I1206 16:41:18.628351 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-85cc9c6459-lnvlp_c4b18e9f-80a6-480e-878b-3acf0a5716a3/proxy-server/0.log" Dec 06 16:41:18 crc kubenswrapper[4813]: I1206 16:41:18.644609 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-85cc9c6459-lnvlp_c4b18e9f-80a6-480e-878b-3acf0a5716a3/proxy-httpd/0.log" Dec 06 16:41:18 crc kubenswrapper[4813]: I1206 16:41:18.735451 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-7pplm_f67c4c64-2c67-46bc-af93-6ca6ac4f4814/swift-ring-rebalance/0.log" Dec 06 16:41:18 crc kubenswrapper[4813]: I1206 16:41:18.887007 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f571761-c8de-485f-89de-514fc0a66a55/account-auditor/0.log" Dec 06 16:41:18 crc kubenswrapper[4813]: I1206 16:41:18.963179 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f571761-c8de-485f-89de-514fc0a66a55/account-reaper/0.log" Dec 06 16:41:19 crc kubenswrapper[4813]: I1206 16:41:19.076093 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f571761-c8de-485f-89de-514fc0a66a55/account-replicator/0.log" Dec 06 16:41:19 crc kubenswrapper[4813]: I1206 16:41:19.183289 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f571761-c8de-485f-89de-514fc0a66a55/account-server/0.log" Dec 06 16:41:19 crc kubenswrapper[4813]: I1206 16:41:19.274038 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f571761-c8de-485f-89de-514fc0a66a55/container-auditor/0.log" Dec 06 16:41:19 crc kubenswrapper[4813]: I1206 16:41:19.344079 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f571761-c8de-485f-89de-514fc0a66a55/container-replicator/0.log" Dec 06 16:41:19 crc kubenswrapper[4813]: I1206 16:41:19.439408 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f571761-c8de-485f-89de-514fc0a66a55/container-server/0.log" Dec 06 16:41:19 crc kubenswrapper[4813]: I1206 16:41:19.477003 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f571761-c8de-485f-89de-514fc0a66a55/container-updater/0.log" Dec 06 16:41:19 crc kubenswrapper[4813]: I1206 16:41:19.532309 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f571761-c8de-485f-89de-514fc0a66a55/object-auditor/0.log" Dec 06 16:41:19 crc kubenswrapper[4813]: I1206 16:41:19.615497 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f571761-c8de-485f-89de-514fc0a66a55/object-expirer/0.log" Dec 06 16:41:19 crc kubenswrapper[4813]: I1206 16:41:19.811222 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f571761-c8de-485f-89de-514fc0a66a55/object-replicator/0.log" Dec 06 16:41:19 crc kubenswrapper[4813]: I1206 16:41:19.823907 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f571761-c8de-485f-89de-514fc0a66a55/object-server/0.log" Dec 06 16:41:19 crc kubenswrapper[4813]: I1206 16:41:19.873841 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f571761-c8de-485f-89de-514fc0a66a55/object-updater/0.log" Dec 06 16:41:20 crc kubenswrapper[4813]: I1206 16:41:20.018322 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f571761-c8de-485f-89de-514fc0a66a55/rsync/0.log" Dec 06 16:41:20 crc kubenswrapper[4813]: I1206 16:41:20.438903 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f571761-c8de-485f-89de-514fc0a66a55/swift-recon-cron/0.log" Dec 06 16:41:20 crc kubenswrapper[4813]: I1206 16:41:20.670900 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt_a1ecf1f7-01c3-456c-8698-1353e0651c24/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:41:20 crc kubenswrapper[4813]: I1206 16:41:20.837846 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_5e4bd588-0ab4-4132-a2e8-1c5ea769f352/tempest-tests-tempest-tests-runner/0.log" Dec 06 16:41:21 crc kubenswrapper[4813]: I1206 16:41:21.035066 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_9899c9cb-fe20-41a1-a200-f70d5b63b314/test-operator-logs-container/0.log" Dec 06 16:41:21 crc kubenswrapper[4813]: I1206 16:41:21.215753 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5_07e3bef8-6975-428b-ac2e-265d2db3d70f/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:41:27 crc kubenswrapper[4813]: I1206 16:41:27.963678 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c/memcached/0.log" Dec 06 16:41:51 crc kubenswrapper[4813]: I1206 16:41:51.987068 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f_89c13d20-79f6-4a15-9c75-cf199679fc7e/util/0.log" Dec 06 16:41:52 crc kubenswrapper[4813]: I1206 16:41:52.321016 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f_89c13d20-79f6-4a15-9c75-cf199679fc7e/util/0.log" Dec 06 16:41:52 crc kubenswrapper[4813]: I1206 16:41:52.322047 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f_89c13d20-79f6-4a15-9c75-cf199679fc7e/pull/0.log" Dec 06 16:41:52 crc kubenswrapper[4813]: I1206 16:41:52.383513 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f_89c13d20-79f6-4a15-9c75-cf199679fc7e/pull/0.log" Dec 06 16:41:53 crc kubenswrapper[4813]: I1206 16:41:53.392027 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f_89c13d20-79f6-4a15-9c75-cf199679fc7e/extract/0.log" Dec 06 16:41:53 crc kubenswrapper[4813]: I1206 16:41:53.455734 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f_89c13d20-79f6-4a15-9c75-cf199679fc7e/util/0.log" Dec 06 16:41:53 crc kubenswrapper[4813]: I1206 16:41:53.496842 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f_89c13d20-79f6-4a15-9c75-cf199679fc7e/pull/0.log" Dec 06 16:41:53 crc kubenswrapper[4813]: I1206 16:41:53.820324 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6dc7dc95b4-6wznz_24d6dd0f-86e6-4b7a-beeb-56025a76ab5a/kube-rbac-proxy/0.log" Dec 06 16:41:53 crc kubenswrapper[4813]: I1206 16:41:53.941544 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-8mt55_a8b0c607-9997-44e8-a3cc-a36e4b2e70d2/kube-rbac-proxy/0.log" Dec 06 16:41:54 crc kubenswrapper[4813]: I1206 16:41:54.025930 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6dc7dc95b4-6wznz_24d6dd0f-86e6-4b7a-beeb-56025a76ab5a/manager/0.log" Dec 06 16:41:54 crc kubenswrapper[4813]: I1206 16:41:54.212613 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-8mt55_a8b0c607-9997-44e8-a3cc-a36e4b2e70d2/manager/0.log" Dec 06 16:41:54 crc kubenswrapper[4813]: I1206 16:41:54.261468 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-4rdzx_218be173-f41f-48b1-8af0-b94bbc483e8b/kube-rbac-proxy/0.log" Dec 06 16:41:54 crc kubenswrapper[4813]: I1206 16:41:54.348591 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-4rdzx_218be173-f41f-48b1-8af0-b94bbc483e8b/manager/0.log" Dec 06 16:41:54 crc kubenswrapper[4813]: I1206 16:41:54.577082 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-f9nfd_3edd7ca8-a3ba-46c5-96c8-77aa9be672b7/kube-rbac-proxy/0.log" Dec 06 16:41:54 crc kubenswrapper[4813]: I1206 16:41:54.585124 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-f9nfd_3edd7ca8-a3ba-46c5-96c8-77aa9be672b7/manager/0.log" Dec 06 16:41:54 crc kubenswrapper[4813]: I1206 16:41:54.751242 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-wxsqx_e33ae276-0e3d-44b5-b970-17e98051c49c/kube-rbac-proxy/0.log" Dec 06 16:41:54 crc kubenswrapper[4813]: I1206 16:41:54.807253 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-wxsqx_e33ae276-0e3d-44b5-b970-17e98051c49c/manager/0.log" Dec 06 16:41:54 crc kubenswrapper[4813]: I1206 16:41:54.930970 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-g4ztj_d67349e5-13fb-4859-ac8d-3e2c6bd67dbb/kube-rbac-proxy/0.log" Dec 06 16:41:55 crc kubenswrapper[4813]: I1206 16:41:55.067057 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-g4ztj_d67349e5-13fb-4859-ac8d-3e2c6bd67dbb/manager/0.log" Dec 06 16:41:55 crc kubenswrapper[4813]: I1206 16:41:55.199341 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-pkvxj_6983648f-60ef-429d-a2fd-de5dc7e6d0ba/kube-rbac-proxy/0.log" Dec 06 16:41:55 crc kubenswrapper[4813]: I1206 16:41:55.391740 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-cxh72_4d2f63b5-c93e-4949-bb79-459dad74dcf8/kube-rbac-proxy/0.log" Dec 06 16:41:55 crc kubenswrapper[4813]: I1206 16:41:55.397458 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-pkvxj_6983648f-60ef-429d-a2fd-de5dc7e6d0ba/manager/0.log" Dec 06 16:41:55 crc kubenswrapper[4813]: I1206 16:41:55.531275 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-cxh72_4d2f63b5-c93e-4949-bb79-459dad74dcf8/manager/0.log" Dec 06 16:41:55 crc kubenswrapper[4813]: I1206 16:41:55.984533 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-cjzht_8870fd51-2141-4cc8-956b-050e849dcbd4/kube-rbac-proxy/0.log" Dec 06 16:41:56 crc kubenswrapper[4813]: I1206 16:41:56.028041 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-cjzht_8870fd51-2141-4cc8-956b-050e849dcbd4/manager/0.log" Dec 06 16:41:56 crc kubenswrapper[4813]: I1206 16:41:56.213322 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-6p6sf_a8763d41-9404-43ec-866b-9244bf2c4af9/kube-rbac-proxy/0.log" Dec 06 16:41:56 crc kubenswrapper[4813]: I1206 16:41:56.213851 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-6p6sf_a8763d41-9404-43ec-866b-9244bf2c4af9/manager/0.log" Dec 06 16:41:56 crc kubenswrapper[4813]: I1206 16:41:56.328630 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-6n2hj_99a0eae7-c990-42cd-a18e-78b81177bbb6/kube-rbac-proxy/0.log" Dec 06 16:41:56 crc kubenswrapper[4813]: I1206 16:41:56.449295 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-6n2hj_99a0eae7-c990-42cd-a18e-78b81177bbb6/manager/0.log" Dec 06 16:41:56 crc kubenswrapper[4813]: I1206 16:41:56.553558 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-6rr2s_b4cd1114-c317-4351-b566-05317dd589f1/kube-rbac-proxy/0.log" Dec 06 16:41:56 crc kubenswrapper[4813]: I1206 16:41:56.553849 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-6rr2s_b4cd1114-c317-4351-b566-05317dd589f1/manager/0.log" Dec 06 16:41:56 crc kubenswrapper[4813]: I1206 16:41:56.668425 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-78fv6_b675acf0-51cc-43d9-b9c6-bd0ee1730d25/kube-rbac-proxy/0.log" Dec 06 16:41:56 crc kubenswrapper[4813]: I1206 16:41:56.799300 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-78fv6_b675acf0-51cc-43d9-b9c6-bd0ee1730d25/manager/0.log" Dec 06 16:41:56 crc kubenswrapper[4813]: I1206 16:41:56.891202 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-g9q96_031a9bd1-298a-4945-a52d-452005c8467e/kube-rbac-proxy/0.log" Dec 06 16:41:56 crc kubenswrapper[4813]: I1206 16:41:56.914295 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-g9q96_031a9bd1-298a-4945-a52d-452005c8467e/manager/0.log" Dec 06 16:41:56 crc kubenswrapper[4813]: I1206 16:41:56.978410 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879fc7jrk_279a2707-03fd-4cbc-9e13-d7a0d2a284db/kube-rbac-proxy/0.log" Dec 06 16:41:57 crc kubenswrapper[4813]: I1206 16:41:57.043865 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879fc7jrk_279a2707-03fd-4cbc-9e13-d7a0d2a284db/manager/0.log" Dec 06 16:41:57 crc kubenswrapper[4813]: I1206 16:41:57.413898 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-cb74cdcd8-r5tfz_d2bd6dc9-ba04-4168-b879-415d9ce263d4/operator/0.log" Dec 06 16:41:57 crc kubenswrapper[4813]: I1206 16:41:57.431948 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-vgjq7_74a02e6c-0417-4c92-9edd-18925cfbf96e/registry-server/0.log" Dec 06 16:41:57 crc kubenswrapper[4813]: I1206 16:41:57.668819 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-2mpwn_c35efae6-9bf9-466d-ac8d-203d8a2c4b97/kube-rbac-proxy/0.log" Dec 06 16:41:57 crc kubenswrapper[4813]: I1206 16:41:57.699738 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-2mpwn_c35efae6-9bf9-466d-ac8d-203d8a2c4b97/manager/0.log" Dec 06 16:41:57 crc kubenswrapper[4813]: I1206 16:41:57.931825 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-rsbqf_67c6b7ea-97cf-4ffa-9f3b-5111d4fbdeba/kube-rbac-proxy/0.log" Dec 06 16:41:57 crc kubenswrapper[4813]: I1206 16:41:57.989836 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-rsbqf_67c6b7ea-97cf-4ffa-9f3b-5111d4fbdeba/manager/0.log" Dec 06 16:41:58 crc kubenswrapper[4813]: I1206 16:41:58.068977 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-9lwhh_00cfc20a-68d6-46be-98b8-83a43c9d11d9/operator/0.log" Dec 06 16:41:58 crc kubenswrapper[4813]: I1206 16:41:58.136002 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5c8695d769-78g26_572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b/manager/0.log" Dec 06 16:41:58 crc kubenswrapper[4813]: I1206 16:41:58.168016 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-x8vkl_b05dafaa-0d81-4b80-ad3e-48da48a1655f/kube-rbac-proxy/0.log" Dec 06 16:41:58 crc kubenswrapper[4813]: I1206 16:41:58.243298 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-x8vkl_b05dafaa-0d81-4b80-ad3e-48da48a1655f/manager/0.log" Dec 06 16:41:58 crc kubenswrapper[4813]: I1206 16:41:58.325653 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-qkrqw_6d54c577-3fdf-48af-9070-301dd3b3eb81/kube-rbac-proxy/0.log" Dec 06 16:41:58 crc kubenswrapper[4813]: I1206 16:41:58.440959 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-qkrqw_6d54c577-3fdf-48af-9070-301dd3b3eb81/manager/0.log" Dec 06 16:41:58 crc kubenswrapper[4813]: I1206 16:41:58.479148 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-cbmwf_28acd95a-e657-46b5-a771-a24b27bdc6fe/kube-rbac-proxy/0.log" Dec 06 16:41:58 crc kubenswrapper[4813]: I1206 16:41:58.521728 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-cbmwf_28acd95a-e657-46b5-a771-a24b27bdc6fe/manager/0.log" Dec 06 16:41:58 crc kubenswrapper[4813]: I1206 16:41:58.640585 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-8zvf5_5a8652b7-8a55-4658-acce-b7573421b94d/kube-rbac-proxy/0.log" Dec 06 16:41:58 crc kubenswrapper[4813]: I1206 16:41:58.670664 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-8zvf5_5a8652b7-8a55-4658-acce-b7573421b94d/manager/0.log" Dec 06 16:42:19 crc kubenswrapper[4813]: I1206 16:42:19.559499 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-mllt2_935ac385-32ee-46c5-bffb-69b89b4af6a9/control-plane-machine-set-operator/0.log" Dec 06 16:42:19 crc kubenswrapper[4813]: I1206 16:42:19.746693 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-dq6hp_0db3a3aa-f872-43c3-ade2-58c55a448fe3/kube-rbac-proxy/0.log" Dec 06 16:42:19 crc kubenswrapper[4813]: I1206 16:42:19.904079 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-dq6hp_0db3a3aa-f872-43c3-ade2-58c55a448fe3/machine-api-operator/0.log" Dec 06 16:42:34 crc kubenswrapper[4813]: I1206 16:42:34.353330 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-gqwzr_9a87a677-1122-46a6-aae7-820387f122d5/cert-manager-controller/0.log" Dec 06 16:42:34 crc kubenswrapper[4813]: I1206 16:42:34.481624 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-6jx97_93a0448f-03c2-4a73-8ebc-b6bc1a5a3cd3/cert-manager-cainjector/0.log" Dec 06 16:42:34 crc kubenswrapper[4813]: I1206 16:42:34.572273 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-4m6ft_b5fd38e1-0a32-486d-839d-e3cce8c150fe/cert-manager-webhook/0.log" Dec 06 16:42:39 crc kubenswrapper[4813]: I1206 16:42:39.153522 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wdvs6"] Dec 06 16:42:39 crc kubenswrapper[4813]: E1206 16:42:39.156647 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a7a9120-6a77-47fd-abfe-4f093e1baf1e" containerName="container-00" Dec 06 16:42:39 crc kubenswrapper[4813]: I1206 16:42:39.156677 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a7a9120-6a77-47fd-abfe-4f093e1baf1e" containerName="container-00" Dec 06 16:42:39 crc kubenswrapper[4813]: I1206 16:42:39.156976 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a7a9120-6a77-47fd-abfe-4f093e1baf1e" containerName="container-00" Dec 06 16:42:39 crc kubenswrapper[4813]: I1206 16:42:39.158320 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wdvs6" Dec 06 16:42:39 crc kubenswrapper[4813]: I1206 16:42:39.186307 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wdvs6"] Dec 06 16:42:39 crc kubenswrapper[4813]: I1206 16:42:39.274440 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9be5ace-25a3-4f84-88bb-8451ff7d808a-catalog-content\") pod \"community-operators-wdvs6\" (UID: \"f9be5ace-25a3-4f84-88bb-8451ff7d808a\") " pod="openshift-marketplace/community-operators-wdvs6" Dec 06 16:42:39 crc kubenswrapper[4813]: I1206 16:42:39.274500 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcxvm\" (UniqueName: \"kubernetes.io/projected/f9be5ace-25a3-4f84-88bb-8451ff7d808a-kube-api-access-dcxvm\") pod \"community-operators-wdvs6\" (UID: \"f9be5ace-25a3-4f84-88bb-8451ff7d808a\") " pod="openshift-marketplace/community-operators-wdvs6" Dec 06 16:42:39 crc kubenswrapper[4813]: I1206 16:42:39.274608 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9be5ace-25a3-4f84-88bb-8451ff7d808a-utilities\") pod \"community-operators-wdvs6\" (UID: \"f9be5ace-25a3-4f84-88bb-8451ff7d808a\") " pod="openshift-marketplace/community-operators-wdvs6" Dec 06 16:42:39 crc kubenswrapper[4813]: I1206 16:42:39.376190 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9be5ace-25a3-4f84-88bb-8451ff7d808a-catalog-content\") pod \"community-operators-wdvs6\" (UID: \"f9be5ace-25a3-4f84-88bb-8451ff7d808a\") " pod="openshift-marketplace/community-operators-wdvs6" Dec 06 16:42:39 crc kubenswrapper[4813]: I1206 16:42:39.376249 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcxvm\" (UniqueName: \"kubernetes.io/projected/f9be5ace-25a3-4f84-88bb-8451ff7d808a-kube-api-access-dcxvm\") pod \"community-operators-wdvs6\" (UID: \"f9be5ace-25a3-4f84-88bb-8451ff7d808a\") " pod="openshift-marketplace/community-operators-wdvs6" Dec 06 16:42:39 crc kubenswrapper[4813]: I1206 16:42:39.376338 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9be5ace-25a3-4f84-88bb-8451ff7d808a-utilities\") pod \"community-operators-wdvs6\" (UID: \"f9be5ace-25a3-4f84-88bb-8451ff7d808a\") " pod="openshift-marketplace/community-operators-wdvs6" Dec 06 16:42:39 crc kubenswrapper[4813]: I1206 16:42:39.376732 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9be5ace-25a3-4f84-88bb-8451ff7d808a-catalog-content\") pod \"community-operators-wdvs6\" (UID: \"f9be5ace-25a3-4f84-88bb-8451ff7d808a\") " pod="openshift-marketplace/community-operators-wdvs6" Dec 06 16:42:39 crc kubenswrapper[4813]: I1206 16:42:39.376786 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9be5ace-25a3-4f84-88bb-8451ff7d808a-utilities\") pod \"community-operators-wdvs6\" (UID: \"f9be5ace-25a3-4f84-88bb-8451ff7d808a\") " pod="openshift-marketplace/community-operators-wdvs6" Dec 06 16:42:39 crc kubenswrapper[4813]: I1206 16:42:39.395354 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcxvm\" (UniqueName: \"kubernetes.io/projected/f9be5ace-25a3-4f84-88bb-8451ff7d808a-kube-api-access-dcxvm\") pod \"community-operators-wdvs6\" (UID: \"f9be5ace-25a3-4f84-88bb-8451ff7d808a\") " pod="openshift-marketplace/community-operators-wdvs6" Dec 06 16:42:39 crc kubenswrapper[4813]: I1206 16:42:39.479298 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wdvs6" Dec 06 16:42:40 crc kubenswrapper[4813]: I1206 16:42:40.013589 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wdvs6"] Dec 06 16:42:40 crc kubenswrapper[4813]: I1206 16:42:40.898709 4813 generic.go:334] "Generic (PLEG): container finished" podID="f9be5ace-25a3-4f84-88bb-8451ff7d808a" containerID="f99c02b37d23e100af2d522a41422a03d7499f01c8aff87964ca2d66b5853511" exitCode=0 Dec 06 16:42:40 crc kubenswrapper[4813]: I1206 16:42:40.899181 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wdvs6" event={"ID":"f9be5ace-25a3-4f84-88bb-8451ff7d808a","Type":"ContainerDied","Data":"f99c02b37d23e100af2d522a41422a03d7499f01c8aff87964ca2d66b5853511"} Dec 06 16:42:40 crc kubenswrapper[4813]: I1206 16:42:40.899224 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wdvs6" event={"ID":"f9be5ace-25a3-4f84-88bb-8451ff7d808a","Type":"ContainerStarted","Data":"439476a2edde522f002700ddd51615c89d6afbe86c35d387062590ca797d7b33"} Dec 06 16:42:40 crc kubenswrapper[4813]: I1206 16:42:40.903193 4813 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 16:42:41 crc kubenswrapper[4813]: I1206 16:42:41.553212 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wmrjg"] Dec 06 16:42:41 crc kubenswrapper[4813]: I1206 16:42:41.555453 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wmrjg" Dec 06 16:42:41 crc kubenswrapper[4813]: I1206 16:42:41.567904 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wmrjg"] Dec 06 16:42:41 crc kubenswrapper[4813]: I1206 16:42:41.717771 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-944t7\" (UniqueName: \"kubernetes.io/projected/ceb0bb9e-e626-4a72-baf3-e1846bb450df-kube-api-access-944t7\") pod \"redhat-marketplace-wmrjg\" (UID: \"ceb0bb9e-e626-4a72-baf3-e1846bb450df\") " pod="openshift-marketplace/redhat-marketplace-wmrjg" Dec 06 16:42:41 crc kubenswrapper[4813]: I1206 16:42:41.717835 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceb0bb9e-e626-4a72-baf3-e1846bb450df-utilities\") pod \"redhat-marketplace-wmrjg\" (UID: \"ceb0bb9e-e626-4a72-baf3-e1846bb450df\") " pod="openshift-marketplace/redhat-marketplace-wmrjg" Dec 06 16:42:41 crc kubenswrapper[4813]: I1206 16:42:41.717863 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceb0bb9e-e626-4a72-baf3-e1846bb450df-catalog-content\") pod \"redhat-marketplace-wmrjg\" (UID: \"ceb0bb9e-e626-4a72-baf3-e1846bb450df\") " pod="openshift-marketplace/redhat-marketplace-wmrjg" Dec 06 16:42:41 crc kubenswrapper[4813]: I1206 16:42:41.819837 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-944t7\" (UniqueName: \"kubernetes.io/projected/ceb0bb9e-e626-4a72-baf3-e1846bb450df-kube-api-access-944t7\") pod \"redhat-marketplace-wmrjg\" (UID: \"ceb0bb9e-e626-4a72-baf3-e1846bb450df\") " pod="openshift-marketplace/redhat-marketplace-wmrjg" Dec 06 16:42:41 crc kubenswrapper[4813]: I1206 16:42:41.819900 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceb0bb9e-e626-4a72-baf3-e1846bb450df-utilities\") pod \"redhat-marketplace-wmrjg\" (UID: \"ceb0bb9e-e626-4a72-baf3-e1846bb450df\") " pod="openshift-marketplace/redhat-marketplace-wmrjg" Dec 06 16:42:41 crc kubenswrapper[4813]: I1206 16:42:41.819924 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceb0bb9e-e626-4a72-baf3-e1846bb450df-catalog-content\") pod \"redhat-marketplace-wmrjg\" (UID: \"ceb0bb9e-e626-4a72-baf3-e1846bb450df\") " pod="openshift-marketplace/redhat-marketplace-wmrjg" Dec 06 16:42:41 crc kubenswrapper[4813]: I1206 16:42:41.820386 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceb0bb9e-e626-4a72-baf3-e1846bb450df-catalog-content\") pod \"redhat-marketplace-wmrjg\" (UID: \"ceb0bb9e-e626-4a72-baf3-e1846bb450df\") " pod="openshift-marketplace/redhat-marketplace-wmrjg" Dec 06 16:42:41 crc kubenswrapper[4813]: I1206 16:42:41.820848 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceb0bb9e-e626-4a72-baf3-e1846bb450df-utilities\") pod \"redhat-marketplace-wmrjg\" (UID: \"ceb0bb9e-e626-4a72-baf3-e1846bb450df\") " pod="openshift-marketplace/redhat-marketplace-wmrjg" Dec 06 16:42:41 crc kubenswrapper[4813]: I1206 16:42:41.844059 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-944t7\" (UniqueName: \"kubernetes.io/projected/ceb0bb9e-e626-4a72-baf3-e1846bb450df-kube-api-access-944t7\") pod \"redhat-marketplace-wmrjg\" (UID: \"ceb0bb9e-e626-4a72-baf3-e1846bb450df\") " pod="openshift-marketplace/redhat-marketplace-wmrjg" Dec 06 16:42:41 crc kubenswrapper[4813]: I1206 16:42:41.876031 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wmrjg" Dec 06 16:42:41 crc kubenswrapper[4813]: I1206 16:42:41.920513 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wdvs6" event={"ID":"f9be5ace-25a3-4f84-88bb-8451ff7d808a","Type":"ContainerStarted","Data":"b9e7cdbd47c0a47740cee97ea498a29f6e6829375decc6905ac02da63058d216"} Dec 06 16:42:42 crc kubenswrapper[4813]: I1206 16:42:42.421683 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wmrjg"] Dec 06 16:42:42 crc kubenswrapper[4813]: W1206 16:42:42.433883 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podceb0bb9e_e626_4a72_baf3_e1846bb450df.slice/crio-9e0d718d4fe7aca85ee67fa8dbdc4345a24eb0f84fae51a086e45dcea56e1e1e WatchSource:0}: Error finding container 9e0d718d4fe7aca85ee67fa8dbdc4345a24eb0f84fae51a086e45dcea56e1e1e: Status 404 returned error can't find the container with id 9e0d718d4fe7aca85ee67fa8dbdc4345a24eb0f84fae51a086e45dcea56e1e1e Dec 06 16:42:42 crc kubenswrapper[4813]: I1206 16:42:42.931169 4813 generic.go:334] "Generic (PLEG): container finished" podID="ceb0bb9e-e626-4a72-baf3-e1846bb450df" containerID="b28a93dc4dd30780d64a7d5211fa190778df097a79415b53a6a6dbe26a68b55e" exitCode=0 Dec 06 16:42:42 crc kubenswrapper[4813]: I1206 16:42:42.931233 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wmrjg" event={"ID":"ceb0bb9e-e626-4a72-baf3-e1846bb450df","Type":"ContainerDied","Data":"b28a93dc4dd30780d64a7d5211fa190778df097a79415b53a6a6dbe26a68b55e"} Dec 06 16:42:42 crc kubenswrapper[4813]: I1206 16:42:42.931330 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wmrjg" event={"ID":"ceb0bb9e-e626-4a72-baf3-e1846bb450df","Type":"ContainerStarted","Data":"9e0d718d4fe7aca85ee67fa8dbdc4345a24eb0f84fae51a086e45dcea56e1e1e"} Dec 06 16:42:42 crc kubenswrapper[4813]: I1206 16:42:42.933911 4813 generic.go:334] "Generic (PLEG): container finished" podID="f9be5ace-25a3-4f84-88bb-8451ff7d808a" containerID="b9e7cdbd47c0a47740cee97ea498a29f6e6829375decc6905ac02da63058d216" exitCode=0 Dec 06 16:42:42 crc kubenswrapper[4813]: I1206 16:42:42.933953 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wdvs6" event={"ID":"f9be5ace-25a3-4f84-88bb-8451ff7d808a","Type":"ContainerDied","Data":"b9e7cdbd47c0a47740cee97ea498a29f6e6829375decc6905ac02da63058d216"} Dec 06 16:42:43 crc kubenswrapper[4813]: I1206 16:42:43.985296 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wdvs6" event={"ID":"f9be5ace-25a3-4f84-88bb-8451ff7d808a","Type":"ContainerStarted","Data":"10dce30a315a21f68924d5399d70ed818339d66e873f4d9e17afdeef0ca132f7"} Dec 06 16:42:43 crc kubenswrapper[4813]: I1206 16:42:43.994038 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wmrjg" event={"ID":"ceb0bb9e-e626-4a72-baf3-e1846bb450df","Type":"ContainerStarted","Data":"fbba582522fa2ca4d99d62e0b6072fade09970467a637654107a2cafb08d6f91"} Dec 06 16:42:44 crc kubenswrapper[4813]: I1206 16:42:44.019797 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wdvs6" podStartSLOduration=2.616159418 podStartE2EDuration="5.019777321s" podCreationTimestamp="2025-12-06 16:42:39 +0000 UTC" firstStartedPulling="2025-12-06 16:42:40.902967747 +0000 UTC m=+3400.793847313" lastFinishedPulling="2025-12-06 16:42:43.30658565 +0000 UTC m=+3403.197465216" observedRunningTime="2025-12-06 16:42:44.007202611 +0000 UTC m=+3403.898082187" watchObservedRunningTime="2025-12-06 16:42:44.019777321 +0000 UTC m=+3403.910656897" Dec 06 16:42:45 crc kubenswrapper[4813]: I1206 16:42:45.004573 4813 generic.go:334] "Generic (PLEG): container finished" podID="ceb0bb9e-e626-4a72-baf3-e1846bb450df" containerID="fbba582522fa2ca4d99d62e0b6072fade09970467a637654107a2cafb08d6f91" exitCode=0 Dec 06 16:42:45 crc kubenswrapper[4813]: I1206 16:42:45.005727 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wmrjg" event={"ID":"ceb0bb9e-e626-4a72-baf3-e1846bb450df","Type":"ContainerDied","Data":"fbba582522fa2ca4d99d62e0b6072fade09970467a637654107a2cafb08d6f91"} Dec 06 16:42:46 crc kubenswrapper[4813]: I1206 16:42:46.014848 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wmrjg" event={"ID":"ceb0bb9e-e626-4a72-baf3-e1846bb450df","Type":"ContainerStarted","Data":"8794a86438b4612ebd4df843f9994d5bb28a65120634e88ceb4ecc09c2f74b61"} Dec 06 16:42:46 crc kubenswrapper[4813]: I1206 16:42:46.034142 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wmrjg" podStartSLOduration=2.54781128 podStartE2EDuration="5.034126224s" podCreationTimestamp="2025-12-06 16:42:41 +0000 UTC" firstStartedPulling="2025-12-06 16:42:42.933564238 +0000 UTC m=+3402.824443834" lastFinishedPulling="2025-12-06 16:42:45.419879202 +0000 UTC m=+3405.310758778" observedRunningTime="2025-12-06 16:42:46.030055277 +0000 UTC m=+3405.920934883" watchObservedRunningTime="2025-12-06 16:42:46.034126224 +0000 UTC m=+3405.925005800" Dec 06 16:42:49 crc kubenswrapper[4813]: I1206 16:42:49.015214 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-r7rqn_2e477e7f-f00a-40a3-b26a-d49ef1f7f41d/nmstate-console-plugin/0.log" Dec 06 16:42:49 crc kubenswrapper[4813]: I1206 16:42:49.178560 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-f4x5w_db7dde13-c533-49ae-bae9-7239ad2b8db6/kube-rbac-proxy/0.log" Dec 06 16:42:49 crc kubenswrapper[4813]: I1206 16:42:49.226926 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-zw99x_4f31e85e-1c68-4627-b047-72de6b05d490/nmstate-handler/0.log" Dec 06 16:42:49 crc kubenswrapper[4813]: I1206 16:42:49.286664 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-f4x5w_db7dde13-c533-49ae-bae9-7239ad2b8db6/nmstate-metrics/0.log" Dec 06 16:42:49 crc kubenswrapper[4813]: I1206 16:42:49.427078 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:42:49 crc kubenswrapper[4813]: I1206 16:42:49.427127 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:42:49 crc kubenswrapper[4813]: I1206 16:42:49.429095 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-bd2wf_215ac015-9856-4c78-ab4d-0035ee29b678/nmstate-operator/0.log" Dec 06 16:42:49 crc kubenswrapper[4813]: I1206 16:42:49.480055 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wdvs6" Dec 06 16:42:49 crc kubenswrapper[4813]: I1206 16:42:49.480097 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wdvs6" Dec 06 16:42:49 crc kubenswrapper[4813]: I1206 16:42:49.495311 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-vc5f4_3b585cd5-6045-46b1-86cf-03536009e643/nmstate-webhook/0.log" Dec 06 16:42:49 crc kubenswrapper[4813]: I1206 16:42:49.528780 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wdvs6" Dec 06 16:42:50 crc kubenswrapper[4813]: I1206 16:42:50.146938 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wdvs6" Dec 06 16:42:50 crc kubenswrapper[4813]: I1206 16:42:50.933668 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wdvs6"] Dec 06 16:42:51 crc kubenswrapper[4813]: I1206 16:42:51.876805 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wmrjg" Dec 06 16:42:51 crc kubenswrapper[4813]: I1206 16:42:51.877801 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wmrjg" Dec 06 16:42:51 crc kubenswrapper[4813]: I1206 16:42:51.936876 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wmrjg" Dec 06 16:42:52 crc kubenswrapper[4813]: I1206 16:42:52.117433 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wdvs6" podUID="f9be5ace-25a3-4f84-88bb-8451ff7d808a" containerName="registry-server" containerID="cri-o://10dce30a315a21f68924d5399d70ed818339d66e873f4d9e17afdeef0ca132f7" gracePeriod=2 Dec 06 16:42:52 crc kubenswrapper[4813]: I1206 16:42:52.171446 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wmrjg" Dec 06 16:42:52 crc kubenswrapper[4813]: I1206 16:42:52.609710 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wdvs6" Dec 06 16:42:52 crc kubenswrapper[4813]: I1206 16:42:52.762221 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcxvm\" (UniqueName: \"kubernetes.io/projected/f9be5ace-25a3-4f84-88bb-8451ff7d808a-kube-api-access-dcxvm\") pod \"f9be5ace-25a3-4f84-88bb-8451ff7d808a\" (UID: \"f9be5ace-25a3-4f84-88bb-8451ff7d808a\") " Dec 06 16:42:52 crc kubenswrapper[4813]: I1206 16:42:52.762412 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9be5ace-25a3-4f84-88bb-8451ff7d808a-catalog-content\") pod \"f9be5ace-25a3-4f84-88bb-8451ff7d808a\" (UID: \"f9be5ace-25a3-4f84-88bb-8451ff7d808a\") " Dec 06 16:42:52 crc kubenswrapper[4813]: I1206 16:42:52.762448 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9be5ace-25a3-4f84-88bb-8451ff7d808a-utilities\") pod \"f9be5ace-25a3-4f84-88bb-8451ff7d808a\" (UID: \"f9be5ace-25a3-4f84-88bb-8451ff7d808a\") " Dec 06 16:42:52 crc kubenswrapper[4813]: I1206 16:42:52.763250 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9be5ace-25a3-4f84-88bb-8451ff7d808a-utilities" (OuterVolumeSpecName: "utilities") pod "f9be5ace-25a3-4f84-88bb-8451ff7d808a" (UID: "f9be5ace-25a3-4f84-88bb-8451ff7d808a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:42:52 crc kubenswrapper[4813]: I1206 16:42:52.768121 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9be5ace-25a3-4f84-88bb-8451ff7d808a-kube-api-access-dcxvm" (OuterVolumeSpecName: "kube-api-access-dcxvm") pod "f9be5ace-25a3-4f84-88bb-8451ff7d808a" (UID: "f9be5ace-25a3-4f84-88bb-8451ff7d808a"). InnerVolumeSpecName "kube-api-access-dcxvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:42:52 crc kubenswrapper[4813]: I1206 16:42:52.816338 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9be5ace-25a3-4f84-88bb-8451ff7d808a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f9be5ace-25a3-4f84-88bb-8451ff7d808a" (UID: "f9be5ace-25a3-4f84-88bb-8451ff7d808a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:42:52 crc kubenswrapper[4813]: I1206 16:42:52.864656 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcxvm\" (UniqueName: \"kubernetes.io/projected/f9be5ace-25a3-4f84-88bb-8451ff7d808a-kube-api-access-dcxvm\") on node \"crc\" DevicePath \"\"" Dec 06 16:42:52 crc kubenswrapper[4813]: I1206 16:42:52.865287 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9be5ace-25a3-4f84-88bb-8451ff7d808a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 16:42:52 crc kubenswrapper[4813]: I1206 16:42:52.865381 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9be5ace-25a3-4f84-88bb-8451ff7d808a-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 16:42:53 crc kubenswrapper[4813]: I1206 16:42:53.128454 4813 generic.go:334] "Generic (PLEG): container finished" podID="f9be5ace-25a3-4f84-88bb-8451ff7d808a" containerID="10dce30a315a21f68924d5399d70ed818339d66e873f4d9e17afdeef0ca132f7" exitCode=0 Dec 06 16:42:53 crc kubenswrapper[4813]: I1206 16:42:53.128546 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wdvs6" Dec 06 16:42:53 crc kubenswrapper[4813]: I1206 16:42:53.128547 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wdvs6" event={"ID":"f9be5ace-25a3-4f84-88bb-8451ff7d808a","Type":"ContainerDied","Data":"10dce30a315a21f68924d5399d70ed818339d66e873f4d9e17afdeef0ca132f7"} Dec 06 16:42:53 crc kubenswrapper[4813]: I1206 16:42:53.129680 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wdvs6" event={"ID":"f9be5ace-25a3-4f84-88bb-8451ff7d808a","Type":"ContainerDied","Data":"439476a2edde522f002700ddd51615c89d6afbe86c35d387062590ca797d7b33"} Dec 06 16:42:53 crc kubenswrapper[4813]: I1206 16:42:53.129715 4813 scope.go:117] "RemoveContainer" containerID="10dce30a315a21f68924d5399d70ed818339d66e873f4d9e17afdeef0ca132f7" Dec 06 16:42:53 crc kubenswrapper[4813]: I1206 16:42:53.162087 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wdvs6"] Dec 06 16:42:53 crc kubenswrapper[4813]: I1206 16:42:53.166059 4813 scope.go:117] "RemoveContainer" containerID="b9e7cdbd47c0a47740cee97ea498a29f6e6829375decc6905ac02da63058d216" Dec 06 16:42:53 crc kubenswrapper[4813]: I1206 16:42:53.176229 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wdvs6"] Dec 06 16:42:53 crc kubenswrapper[4813]: I1206 16:42:53.196235 4813 scope.go:117] "RemoveContainer" containerID="f99c02b37d23e100af2d522a41422a03d7499f01c8aff87964ca2d66b5853511" Dec 06 16:42:53 crc kubenswrapper[4813]: I1206 16:42:53.251011 4813 scope.go:117] "RemoveContainer" containerID="10dce30a315a21f68924d5399d70ed818339d66e873f4d9e17afdeef0ca132f7" Dec 06 16:42:53 crc kubenswrapper[4813]: E1206 16:42:53.251951 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10dce30a315a21f68924d5399d70ed818339d66e873f4d9e17afdeef0ca132f7\": container with ID starting with 10dce30a315a21f68924d5399d70ed818339d66e873f4d9e17afdeef0ca132f7 not found: ID does not exist" containerID="10dce30a315a21f68924d5399d70ed818339d66e873f4d9e17afdeef0ca132f7" Dec 06 16:42:53 crc kubenswrapper[4813]: I1206 16:42:53.251993 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10dce30a315a21f68924d5399d70ed818339d66e873f4d9e17afdeef0ca132f7"} err="failed to get container status \"10dce30a315a21f68924d5399d70ed818339d66e873f4d9e17afdeef0ca132f7\": rpc error: code = NotFound desc = could not find container \"10dce30a315a21f68924d5399d70ed818339d66e873f4d9e17afdeef0ca132f7\": container with ID starting with 10dce30a315a21f68924d5399d70ed818339d66e873f4d9e17afdeef0ca132f7 not found: ID does not exist" Dec 06 16:42:53 crc kubenswrapper[4813]: I1206 16:42:53.252045 4813 scope.go:117] "RemoveContainer" containerID="b9e7cdbd47c0a47740cee97ea498a29f6e6829375decc6905ac02da63058d216" Dec 06 16:42:53 crc kubenswrapper[4813]: E1206 16:42:53.255652 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9e7cdbd47c0a47740cee97ea498a29f6e6829375decc6905ac02da63058d216\": container with ID starting with b9e7cdbd47c0a47740cee97ea498a29f6e6829375decc6905ac02da63058d216 not found: ID does not exist" containerID="b9e7cdbd47c0a47740cee97ea498a29f6e6829375decc6905ac02da63058d216" Dec 06 16:42:53 crc kubenswrapper[4813]: I1206 16:42:53.255695 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9e7cdbd47c0a47740cee97ea498a29f6e6829375decc6905ac02da63058d216"} err="failed to get container status \"b9e7cdbd47c0a47740cee97ea498a29f6e6829375decc6905ac02da63058d216\": rpc error: code = NotFound desc = could not find container \"b9e7cdbd47c0a47740cee97ea498a29f6e6829375decc6905ac02da63058d216\": container with ID starting with b9e7cdbd47c0a47740cee97ea498a29f6e6829375decc6905ac02da63058d216 not found: ID does not exist" Dec 06 16:42:53 crc kubenswrapper[4813]: I1206 16:42:53.255716 4813 scope.go:117] "RemoveContainer" containerID="f99c02b37d23e100af2d522a41422a03d7499f01c8aff87964ca2d66b5853511" Dec 06 16:42:53 crc kubenswrapper[4813]: E1206 16:42:53.256082 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f99c02b37d23e100af2d522a41422a03d7499f01c8aff87964ca2d66b5853511\": container with ID starting with f99c02b37d23e100af2d522a41422a03d7499f01c8aff87964ca2d66b5853511 not found: ID does not exist" containerID="f99c02b37d23e100af2d522a41422a03d7499f01c8aff87964ca2d66b5853511" Dec 06 16:42:53 crc kubenswrapper[4813]: I1206 16:42:53.256114 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f99c02b37d23e100af2d522a41422a03d7499f01c8aff87964ca2d66b5853511"} err="failed to get container status \"f99c02b37d23e100af2d522a41422a03d7499f01c8aff87964ca2d66b5853511\": rpc error: code = NotFound desc = could not find container \"f99c02b37d23e100af2d522a41422a03d7499f01c8aff87964ca2d66b5853511\": container with ID starting with f99c02b37d23e100af2d522a41422a03d7499f01c8aff87964ca2d66b5853511 not found: ID does not exist" Dec 06 16:42:54 crc kubenswrapper[4813]: I1206 16:42:54.336593 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wmrjg"] Dec 06 16:42:54 crc kubenswrapper[4813]: I1206 16:42:54.497078 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9be5ace-25a3-4f84-88bb-8451ff7d808a" path="/var/lib/kubelet/pods/f9be5ace-25a3-4f84-88bb-8451ff7d808a/volumes" Dec 06 16:42:55 crc kubenswrapper[4813]: I1206 16:42:55.149495 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wmrjg" podUID="ceb0bb9e-e626-4a72-baf3-e1846bb450df" containerName="registry-server" containerID="cri-o://8794a86438b4612ebd4df843f9994d5bb28a65120634e88ceb4ecc09c2f74b61" gracePeriod=2 Dec 06 16:42:55 crc kubenswrapper[4813]: I1206 16:42:55.625351 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wmrjg" Dec 06 16:42:55 crc kubenswrapper[4813]: I1206 16:42:55.726368 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceb0bb9e-e626-4a72-baf3-e1846bb450df-utilities\") pod \"ceb0bb9e-e626-4a72-baf3-e1846bb450df\" (UID: \"ceb0bb9e-e626-4a72-baf3-e1846bb450df\") " Dec 06 16:42:55 crc kubenswrapper[4813]: I1206 16:42:55.726740 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceb0bb9e-e626-4a72-baf3-e1846bb450df-catalog-content\") pod \"ceb0bb9e-e626-4a72-baf3-e1846bb450df\" (UID: \"ceb0bb9e-e626-4a72-baf3-e1846bb450df\") " Dec 06 16:42:55 crc kubenswrapper[4813]: I1206 16:42:55.726908 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-944t7\" (UniqueName: \"kubernetes.io/projected/ceb0bb9e-e626-4a72-baf3-e1846bb450df-kube-api-access-944t7\") pod \"ceb0bb9e-e626-4a72-baf3-e1846bb450df\" (UID: \"ceb0bb9e-e626-4a72-baf3-e1846bb450df\") " Dec 06 16:42:55 crc kubenswrapper[4813]: I1206 16:42:55.726982 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ceb0bb9e-e626-4a72-baf3-e1846bb450df-utilities" (OuterVolumeSpecName: "utilities") pod "ceb0bb9e-e626-4a72-baf3-e1846bb450df" (UID: "ceb0bb9e-e626-4a72-baf3-e1846bb450df"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:42:55 crc kubenswrapper[4813]: I1206 16:42:55.727422 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceb0bb9e-e626-4a72-baf3-e1846bb450df-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 16:42:55 crc kubenswrapper[4813]: I1206 16:42:55.740650 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ceb0bb9e-e626-4a72-baf3-e1846bb450df-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ceb0bb9e-e626-4a72-baf3-e1846bb450df" (UID: "ceb0bb9e-e626-4a72-baf3-e1846bb450df"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:42:55 crc kubenswrapper[4813]: I1206 16:42:55.749437 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ceb0bb9e-e626-4a72-baf3-e1846bb450df-kube-api-access-944t7" (OuterVolumeSpecName: "kube-api-access-944t7") pod "ceb0bb9e-e626-4a72-baf3-e1846bb450df" (UID: "ceb0bb9e-e626-4a72-baf3-e1846bb450df"). InnerVolumeSpecName "kube-api-access-944t7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:42:55 crc kubenswrapper[4813]: I1206 16:42:55.828821 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-944t7\" (UniqueName: \"kubernetes.io/projected/ceb0bb9e-e626-4a72-baf3-e1846bb450df-kube-api-access-944t7\") on node \"crc\" DevicePath \"\"" Dec 06 16:42:55 crc kubenswrapper[4813]: I1206 16:42:55.828863 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceb0bb9e-e626-4a72-baf3-e1846bb450df-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 16:42:56 crc kubenswrapper[4813]: I1206 16:42:56.161147 4813 generic.go:334] "Generic (PLEG): container finished" podID="ceb0bb9e-e626-4a72-baf3-e1846bb450df" containerID="8794a86438b4612ebd4df843f9994d5bb28a65120634e88ceb4ecc09c2f74b61" exitCode=0 Dec 06 16:42:56 crc kubenswrapper[4813]: I1206 16:42:56.161216 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wmrjg" Dec 06 16:42:56 crc kubenswrapper[4813]: I1206 16:42:56.161195 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wmrjg" event={"ID":"ceb0bb9e-e626-4a72-baf3-e1846bb450df","Type":"ContainerDied","Data":"8794a86438b4612ebd4df843f9994d5bb28a65120634e88ceb4ecc09c2f74b61"} Dec 06 16:42:56 crc kubenswrapper[4813]: I1206 16:42:56.161316 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wmrjg" event={"ID":"ceb0bb9e-e626-4a72-baf3-e1846bb450df","Type":"ContainerDied","Data":"9e0d718d4fe7aca85ee67fa8dbdc4345a24eb0f84fae51a086e45dcea56e1e1e"} Dec 06 16:42:56 crc kubenswrapper[4813]: I1206 16:42:56.161340 4813 scope.go:117] "RemoveContainer" containerID="8794a86438b4612ebd4df843f9994d5bb28a65120634e88ceb4ecc09c2f74b61" Dec 06 16:42:56 crc kubenswrapper[4813]: I1206 16:42:56.183977 4813 scope.go:117] "RemoveContainer" containerID="fbba582522fa2ca4d99d62e0b6072fade09970467a637654107a2cafb08d6f91" Dec 06 16:42:56 crc kubenswrapper[4813]: I1206 16:42:56.206580 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wmrjg"] Dec 06 16:42:56 crc kubenswrapper[4813]: I1206 16:42:56.217022 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wmrjg"] Dec 06 16:42:56 crc kubenswrapper[4813]: I1206 16:42:56.230382 4813 scope.go:117] "RemoveContainer" containerID="b28a93dc4dd30780d64a7d5211fa190778df097a79415b53a6a6dbe26a68b55e" Dec 06 16:42:56 crc kubenswrapper[4813]: I1206 16:42:56.273176 4813 scope.go:117] "RemoveContainer" containerID="8794a86438b4612ebd4df843f9994d5bb28a65120634e88ceb4ecc09c2f74b61" Dec 06 16:42:56 crc kubenswrapper[4813]: E1206 16:42:56.273928 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8794a86438b4612ebd4df843f9994d5bb28a65120634e88ceb4ecc09c2f74b61\": container with ID starting with 8794a86438b4612ebd4df843f9994d5bb28a65120634e88ceb4ecc09c2f74b61 not found: ID does not exist" containerID="8794a86438b4612ebd4df843f9994d5bb28a65120634e88ceb4ecc09c2f74b61" Dec 06 16:42:56 crc kubenswrapper[4813]: I1206 16:42:56.273987 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8794a86438b4612ebd4df843f9994d5bb28a65120634e88ceb4ecc09c2f74b61"} err="failed to get container status \"8794a86438b4612ebd4df843f9994d5bb28a65120634e88ceb4ecc09c2f74b61\": rpc error: code = NotFound desc = could not find container \"8794a86438b4612ebd4df843f9994d5bb28a65120634e88ceb4ecc09c2f74b61\": container with ID starting with 8794a86438b4612ebd4df843f9994d5bb28a65120634e88ceb4ecc09c2f74b61 not found: ID does not exist" Dec 06 16:42:56 crc kubenswrapper[4813]: I1206 16:42:56.274024 4813 scope.go:117] "RemoveContainer" containerID="fbba582522fa2ca4d99d62e0b6072fade09970467a637654107a2cafb08d6f91" Dec 06 16:42:56 crc kubenswrapper[4813]: E1206 16:42:56.274364 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbba582522fa2ca4d99d62e0b6072fade09970467a637654107a2cafb08d6f91\": container with ID starting with fbba582522fa2ca4d99d62e0b6072fade09970467a637654107a2cafb08d6f91 not found: ID does not exist" containerID="fbba582522fa2ca4d99d62e0b6072fade09970467a637654107a2cafb08d6f91" Dec 06 16:42:56 crc kubenswrapper[4813]: I1206 16:42:56.274393 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbba582522fa2ca4d99d62e0b6072fade09970467a637654107a2cafb08d6f91"} err="failed to get container status \"fbba582522fa2ca4d99d62e0b6072fade09970467a637654107a2cafb08d6f91\": rpc error: code = NotFound desc = could not find container \"fbba582522fa2ca4d99d62e0b6072fade09970467a637654107a2cafb08d6f91\": container with ID starting with fbba582522fa2ca4d99d62e0b6072fade09970467a637654107a2cafb08d6f91 not found: ID does not exist" Dec 06 16:42:56 crc kubenswrapper[4813]: I1206 16:42:56.274415 4813 scope.go:117] "RemoveContainer" containerID="b28a93dc4dd30780d64a7d5211fa190778df097a79415b53a6a6dbe26a68b55e" Dec 06 16:42:56 crc kubenswrapper[4813]: E1206 16:42:56.274653 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b28a93dc4dd30780d64a7d5211fa190778df097a79415b53a6a6dbe26a68b55e\": container with ID starting with b28a93dc4dd30780d64a7d5211fa190778df097a79415b53a6a6dbe26a68b55e not found: ID does not exist" containerID="b28a93dc4dd30780d64a7d5211fa190778df097a79415b53a6a6dbe26a68b55e" Dec 06 16:42:56 crc kubenswrapper[4813]: I1206 16:42:56.274677 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b28a93dc4dd30780d64a7d5211fa190778df097a79415b53a6a6dbe26a68b55e"} err="failed to get container status \"b28a93dc4dd30780d64a7d5211fa190778df097a79415b53a6a6dbe26a68b55e\": rpc error: code = NotFound desc = could not find container \"b28a93dc4dd30780d64a7d5211fa190778df097a79415b53a6a6dbe26a68b55e\": container with ID starting with b28a93dc4dd30780d64a7d5211fa190778df097a79415b53a6a6dbe26a68b55e not found: ID does not exist" Dec 06 16:42:56 crc kubenswrapper[4813]: I1206 16:42:56.496602 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ceb0bb9e-e626-4a72-baf3-e1846bb450df" path="/var/lib/kubelet/pods/ceb0bb9e-e626-4a72-baf3-e1846bb450df/volumes" Dec 06 16:43:05 crc kubenswrapper[4813]: I1206 16:43:05.826039 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-mfvl4_a8a9650e-6412-4d98-bdf5-6b57f014eb64/kube-rbac-proxy/0.log" Dec 06 16:43:05 crc kubenswrapper[4813]: I1206 16:43:05.950421 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-mfvl4_a8a9650e-6412-4d98-bdf5-6b57f014eb64/controller/0.log" Dec 06 16:43:06 crc kubenswrapper[4813]: I1206 16:43:06.077501 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/cp-frr-files/0.log" Dec 06 16:43:06 crc kubenswrapper[4813]: I1206 16:43:06.219837 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/cp-frr-files/0.log" Dec 06 16:43:06 crc kubenswrapper[4813]: I1206 16:43:06.225612 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/cp-reloader/0.log" Dec 06 16:43:06 crc kubenswrapper[4813]: I1206 16:43:06.300998 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/cp-metrics/0.log" Dec 06 16:43:06 crc kubenswrapper[4813]: I1206 16:43:06.303592 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/cp-reloader/0.log" Dec 06 16:43:06 crc kubenswrapper[4813]: I1206 16:43:06.478333 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/cp-reloader/0.log" Dec 06 16:43:06 crc kubenswrapper[4813]: I1206 16:43:06.527868 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/cp-metrics/0.log" Dec 06 16:43:06 crc kubenswrapper[4813]: I1206 16:43:06.543592 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/cp-frr-files/0.log" Dec 06 16:43:06 crc kubenswrapper[4813]: I1206 16:43:06.579211 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/cp-metrics/0.log" Dec 06 16:43:06 crc kubenswrapper[4813]: I1206 16:43:06.738849 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/cp-metrics/0.log" Dec 06 16:43:06 crc kubenswrapper[4813]: I1206 16:43:06.752200 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/controller/0.log" Dec 06 16:43:06 crc kubenswrapper[4813]: I1206 16:43:06.768150 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/cp-frr-files/0.log" Dec 06 16:43:06 crc kubenswrapper[4813]: I1206 16:43:06.788872 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/cp-reloader/0.log" Dec 06 16:43:06 crc kubenswrapper[4813]: I1206 16:43:06.930015 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/frr-metrics/0.log" Dec 06 16:43:06 crc kubenswrapper[4813]: I1206 16:43:06.953085 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/kube-rbac-proxy/0.log" Dec 06 16:43:07 crc kubenswrapper[4813]: I1206 16:43:07.120446 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/kube-rbac-proxy-frr/0.log" Dec 06 16:43:07 crc kubenswrapper[4813]: I1206 16:43:07.156684 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/reloader/0.log" Dec 06 16:43:07 crc kubenswrapper[4813]: I1206 16:43:07.404123 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-wsvkk_a5c26c26-14a5-406c-882b-2950fd2855ed/frr-k8s-webhook-server/0.log" Dec 06 16:43:07 crc kubenswrapper[4813]: I1206 16:43:07.559063 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6977db5dfd-44msc_e7537632-d133-4c59-a167-e59127ef6fab/manager/0.log" Dec 06 16:43:07 crc kubenswrapper[4813]: I1206 16:43:07.751750 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-65d646cc59-hk6wv_2fcdfd68-9b08-423f-80ec-53032016fdfe/webhook-server/0.log" Dec 06 16:43:07 crc kubenswrapper[4813]: I1206 16:43:07.886626 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/frr/0.log" Dec 06 16:43:07 crc kubenswrapper[4813]: I1206 16:43:07.977917 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-ccll7_45152d05-2dee-4bba-bc56-2e378d0edb9a/kube-rbac-proxy/0.log" Dec 06 16:43:08 crc kubenswrapper[4813]: I1206 16:43:08.226650 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-ccll7_45152d05-2dee-4bba-bc56-2e378d0edb9a/speaker/0.log" Dec 06 16:43:19 crc kubenswrapper[4813]: I1206 16:43:19.427955 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:43:19 crc kubenswrapper[4813]: I1206 16:43:19.428606 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:43:21 crc kubenswrapper[4813]: I1206 16:43:21.137739 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx_4cc38736-85b3-4e2d-87fe-604eebf52317/util/0.log" Dec 06 16:43:21 crc kubenswrapper[4813]: I1206 16:43:21.317660 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx_4cc38736-85b3-4e2d-87fe-604eebf52317/pull/0.log" Dec 06 16:43:21 crc kubenswrapper[4813]: I1206 16:43:21.322996 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx_4cc38736-85b3-4e2d-87fe-604eebf52317/util/0.log" Dec 06 16:43:21 crc kubenswrapper[4813]: I1206 16:43:21.353404 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx_4cc38736-85b3-4e2d-87fe-604eebf52317/pull/0.log" Dec 06 16:43:21 crc kubenswrapper[4813]: I1206 16:43:21.510478 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx_4cc38736-85b3-4e2d-87fe-604eebf52317/extract/0.log" Dec 06 16:43:21 crc kubenswrapper[4813]: I1206 16:43:21.520913 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx_4cc38736-85b3-4e2d-87fe-604eebf52317/util/0.log" Dec 06 16:43:21 crc kubenswrapper[4813]: I1206 16:43:21.521722 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx_4cc38736-85b3-4e2d-87fe-604eebf52317/pull/0.log" Dec 06 16:43:21 crc kubenswrapper[4813]: I1206 16:43:21.726837 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5_6f21e1e0-2a67-4f45-93f4-e84a5bba3007/util/0.log" Dec 06 16:43:21 crc kubenswrapper[4813]: I1206 16:43:21.849863 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5_6f21e1e0-2a67-4f45-93f4-e84a5bba3007/util/0.log" Dec 06 16:43:21 crc kubenswrapper[4813]: I1206 16:43:21.882067 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5_6f21e1e0-2a67-4f45-93f4-e84a5bba3007/pull/0.log" Dec 06 16:43:21 crc kubenswrapper[4813]: I1206 16:43:21.934953 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5_6f21e1e0-2a67-4f45-93f4-e84a5bba3007/pull/0.log" Dec 06 16:43:22 crc kubenswrapper[4813]: I1206 16:43:22.049794 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5_6f21e1e0-2a67-4f45-93f4-e84a5bba3007/pull/0.log" Dec 06 16:43:22 crc kubenswrapper[4813]: I1206 16:43:22.060401 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5_6f21e1e0-2a67-4f45-93f4-e84a5bba3007/util/0.log" Dec 06 16:43:22 crc kubenswrapper[4813]: I1206 16:43:22.109169 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5_6f21e1e0-2a67-4f45-93f4-e84a5bba3007/extract/0.log" Dec 06 16:43:22 crc kubenswrapper[4813]: I1206 16:43:22.268685 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ws47p_abeb6c67-8702-4211-9aab-7cdf25c0d078/extract-utilities/0.log" Dec 06 16:43:22 crc kubenswrapper[4813]: I1206 16:43:22.490333 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ws47p_abeb6c67-8702-4211-9aab-7cdf25c0d078/extract-content/0.log" Dec 06 16:43:22 crc kubenswrapper[4813]: I1206 16:43:22.500041 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ws47p_abeb6c67-8702-4211-9aab-7cdf25c0d078/extract-utilities/0.log" Dec 06 16:43:22 crc kubenswrapper[4813]: I1206 16:43:22.534310 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ws47p_abeb6c67-8702-4211-9aab-7cdf25c0d078/extract-content/0.log" Dec 06 16:43:22 crc kubenswrapper[4813]: I1206 16:43:22.698537 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ws47p_abeb6c67-8702-4211-9aab-7cdf25c0d078/extract-utilities/0.log" Dec 06 16:43:22 crc kubenswrapper[4813]: I1206 16:43:22.710753 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ws47p_abeb6c67-8702-4211-9aab-7cdf25c0d078/extract-content/0.log" Dec 06 16:43:22 crc kubenswrapper[4813]: I1206 16:43:22.908722 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-crqpg_4b8fe11e-abdb-42d7-9d88-7d463a808b4b/extract-utilities/0.log" Dec 06 16:43:23 crc kubenswrapper[4813]: I1206 16:43:23.165125 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ws47p_abeb6c67-8702-4211-9aab-7cdf25c0d078/registry-server/0.log" Dec 06 16:43:23 crc kubenswrapper[4813]: I1206 16:43:23.172493 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-crqpg_4b8fe11e-abdb-42d7-9d88-7d463a808b4b/extract-content/0.log" Dec 06 16:43:23 crc kubenswrapper[4813]: I1206 16:43:23.243589 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-crqpg_4b8fe11e-abdb-42d7-9d88-7d463a808b4b/extract-utilities/0.log" Dec 06 16:43:23 crc kubenswrapper[4813]: I1206 16:43:23.243870 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-crqpg_4b8fe11e-abdb-42d7-9d88-7d463a808b4b/extract-content/0.log" Dec 06 16:43:23 crc kubenswrapper[4813]: I1206 16:43:23.400529 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-crqpg_4b8fe11e-abdb-42d7-9d88-7d463a808b4b/extract-utilities/0.log" Dec 06 16:43:23 crc kubenswrapper[4813]: I1206 16:43:23.422346 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-crqpg_4b8fe11e-abdb-42d7-9d88-7d463a808b4b/extract-content/0.log" Dec 06 16:43:23 crc kubenswrapper[4813]: I1206 16:43:23.665136 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-v6fmp_2660f707-b9cc-4c4f-b42a-9aa6699ef326/marketplace-operator/0.log" Dec 06 16:43:23 crc kubenswrapper[4813]: I1206 16:43:23.822210 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xwq7l_ca46ce83-a423-478a-9d44-d45a0ef28ef3/extract-utilities/0.log" Dec 06 16:43:23 crc kubenswrapper[4813]: I1206 16:43:23.840344 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-crqpg_4b8fe11e-abdb-42d7-9d88-7d463a808b4b/registry-server/0.log" Dec 06 16:43:24 crc kubenswrapper[4813]: I1206 16:43:24.080708 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xwq7l_ca46ce83-a423-478a-9d44-d45a0ef28ef3/extract-utilities/0.log" Dec 06 16:43:24 crc kubenswrapper[4813]: I1206 16:43:24.086780 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xwq7l_ca46ce83-a423-478a-9d44-d45a0ef28ef3/extract-content/0.log" Dec 06 16:43:24 crc kubenswrapper[4813]: I1206 16:43:24.115307 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xwq7l_ca46ce83-a423-478a-9d44-d45a0ef28ef3/extract-content/0.log" Dec 06 16:43:24 crc kubenswrapper[4813]: I1206 16:43:24.254747 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xwq7l_ca46ce83-a423-478a-9d44-d45a0ef28ef3/extract-content/0.log" Dec 06 16:43:24 crc kubenswrapper[4813]: I1206 16:43:24.349553 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xwq7l_ca46ce83-a423-478a-9d44-d45a0ef28ef3/extract-utilities/0.log" Dec 06 16:43:24 crc kubenswrapper[4813]: I1206 16:43:24.369335 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xwq7l_ca46ce83-a423-478a-9d44-d45a0ef28ef3/registry-server/0.log" Dec 06 16:43:24 crc kubenswrapper[4813]: I1206 16:43:24.480796 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-28zj9_46d12276-3ced-4cbe-a060-704f83fcf9a8/extract-utilities/0.log" Dec 06 16:43:24 crc kubenswrapper[4813]: I1206 16:43:24.654963 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-28zj9_46d12276-3ced-4cbe-a060-704f83fcf9a8/extract-content/0.log" Dec 06 16:43:24 crc kubenswrapper[4813]: I1206 16:43:24.680343 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-28zj9_46d12276-3ced-4cbe-a060-704f83fcf9a8/extract-content/0.log" Dec 06 16:43:24 crc kubenswrapper[4813]: I1206 16:43:24.697037 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-28zj9_46d12276-3ced-4cbe-a060-704f83fcf9a8/extract-utilities/0.log" Dec 06 16:43:24 crc kubenswrapper[4813]: I1206 16:43:24.869061 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-28zj9_46d12276-3ced-4cbe-a060-704f83fcf9a8/extract-content/0.log" Dec 06 16:43:24 crc kubenswrapper[4813]: I1206 16:43:24.889035 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-28zj9_46d12276-3ced-4cbe-a060-704f83fcf9a8/extract-utilities/0.log" Dec 06 16:43:25 crc kubenswrapper[4813]: I1206 16:43:25.435894 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-28zj9_46d12276-3ced-4cbe-a060-704f83fcf9a8/registry-server/0.log" Dec 06 16:43:49 crc kubenswrapper[4813]: I1206 16:43:49.427441 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:43:49 crc kubenswrapper[4813]: I1206 16:43:49.427909 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:43:49 crc kubenswrapper[4813]: I1206 16:43:49.427961 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" Dec 06 16:43:49 crc kubenswrapper[4813]: I1206 16:43:49.428720 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"89167fab7a057df4dbc2e1e5aa930f5ed89747de01e89ff0bb679d87ca1d7a24"} pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 16:43:49 crc kubenswrapper[4813]: I1206 16:43:49.428767 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" containerID="cri-o://89167fab7a057df4dbc2e1e5aa930f5ed89747de01e89ff0bb679d87ca1d7a24" gracePeriod=600 Dec 06 16:43:49 crc kubenswrapper[4813]: I1206 16:43:49.677197 4813 generic.go:334] "Generic (PLEG): container finished" podID="d88e8bae-c055-4c55-b548-f621ff96de06" containerID="89167fab7a057df4dbc2e1e5aa930f5ed89747de01e89ff0bb679d87ca1d7a24" exitCode=0 Dec 06 16:43:49 crc kubenswrapper[4813]: I1206 16:43:49.677298 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerDied","Data":"89167fab7a057df4dbc2e1e5aa930f5ed89747de01e89ff0bb679d87ca1d7a24"} Dec 06 16:43:49 crc kubenswrapper[4813]: I1206 16:43:49.677556 4813 scope.go:117] "RemoveContainer" containerID="8441b376fae9405da00e04a7e75ce187d0f25185f6fc0f6fc4224d7561e835bd" Dec 06 16:43:50 crc kubenswrapper[4813]: I1206 16:43:50.692572 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerStarted","Data":"969c7429b9faa2222909b4bad4ef827351b31f2cde6d6ab68e5af156f1ab0dfc"} Dec 06 16:45:00 crc kubenswrapper[4813]: I1206 16:45:00.205685 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417325-p9ml7"] Dec 06 16:45:00 crc kubenswrapper[4813]: E1206 16:45:00.206872 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9be5ace-25a3-4f84-88bb-8451ff7d808a" containerName="extract-utilities" Dec 06 16:45:00 crc kubenswrapper[4813]: I1206 16:45:00.206895 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9be5ace-25a3-4f84-88bb-8451ff7d808a" containerName="extract-utilities" Dec 06 16:45:00 crc kubenswrapper[4813]: E1206 16:45:00.206922 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceb0bb9e-e626-4a72-baf3-e1846bb450df" containerName="registry-server" Dec 06 16:45:00 crc kubenswrapper[4813]: I1206 16:45:00.206936 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceb0bb9e-e626-4a72-baf3-e1846bb450df" containerName="registry-server" Dec 06 16:45:00 crc kubenswrapper[4813]: E1206 16:45:00.206973 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9be5ace-25a3-4f84-88bb-8451ff7d808a" containerName="registry-server" Dec 06 16:45:00 crc kubenswrapper[4813]: I1206 16:45:00.206986 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9be5ace-25a3-4f84-88bb-8451ff7d808a" containerName="registry-server" Dec 06 16:45:00 crc kubenswrapper[4813]: E1206 16:45:00.207019 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceb0bb9e-e626-4a72-baf3-e1846bb450df" containerName="extract-utilities" Dec 06 16:45:00 crc kubenswrapper[4813]: I1206 16:45:00.207033 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceb0bb9e-e626-4a72-baf3-e1846bb450df" containerName="extract-utilities" Dec 06 16:45:00 crc kubenswrapper[4813]: E1206 16:45:00.207066 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceb0bb9e-e626-4a72-baf3-e1846bb450df" containerName="extract-content" Dec 06 16:45:00 crc kubenswrapper[4813]: I1206 16:45:00.207078 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceb0bb9e-e626-4a72-baf3-e1846bb450df" containerName="extract-content" Dec 06 16:45:00 crc kubenswrapper[4813]: E1206 16:45:00.207103 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9be5ace-25a3-4f84-88bb-8451ff7d808a" containerName="extract-content" Dec 06 16:45:00 crc kubenswrapper[4813]: I1206 16:45:00.207115 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9be5ace-25a3-4f84-88bb-8451ff7d808a" containerName="extract-content" Dec 06 16:45:00 crc kubenswrapper[4813]: I1206 16:45:00.207462 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="ceb0bb9e-e626-4a72-baf3-e1846bb450df" containerName="registry-server" Dec 06 16:45:00 crc kubenswrapper[4813]: I1206 16:45:00.207522 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9be5ace-25a3-4f84-88bb-8451ff7d808a" containerName="registry-server" Dec 06 16:45:00 crc kubenswrapper[4813]: I1206 16:45:00.208555 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417325-p9ml7" Dec 06 16:45:00 crc kubenswrapper[4813]: I1206 16:45:00.216861 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 16:45:00 crc kubenswrapper[4813]: I1206 16:45:00.217352 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 16:45:00 crc kubenswrapper[4813]: I1206 16:45:00.225940 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417325-p9ml7"] Dec 06 16:45:00 crc kubenswrapper[4813]: I1206 16:45:00.246802 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/be550258-23b4-48ba-b6e3-9d92a54b582f-config-volume\") pod \"collect-profiles-29417325-p9ml7\" (UID: \"be550258-23b4-48ba-b6e3-9d92a54b582f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417325-p9ml7" Dec 06 16:45:00 crc kubenswrapper[4813]: I1206 16:45:00.246868 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/be550258-23b4-48ba-b6e3-9d92a54b582f-secret-volume\") pod \"collect-profiles-29417325-p9ml7\" (UID: \"be550258-23b4-48ba-b6e3-9d92a54b582f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417325-p9ml7" Dec 06 16:45:00 crc kubenswrapper[4813]: I1206 16:45:00.246919 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mz6j\" (UniqueName: \"kubernetes.io/projected/be550258-23b4-48ba-b6e3-9d92a54b582f-kube-api-access-8mz6j\") pod \"collect-profiles-29417325-p9ml7\" (UID: \"be550258-23b4-48ba-b6e3-9d92a54b582f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417325-p9ml7" Dec 06 16:45:00 crc kubenswrapper[4813]: I1206 16:45:00.349866 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/be550258-23b4-48ba-b6e3-9d92a54b582f-config-volume\") pod \"collect-profiles-29417325-p9ml7\" (UID: \"be550258-23b4-48ba-b6e3-9d92a54b582f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417325-p9ml7" Dec 06 16:45:00 crc kubenswrapper[4813]: I1206 16:45:00.350372 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/be550258-23b4-48ba-b6e3-9d92a54b582f-secret-volume\") pod \"collect-profiles-29417325-p9ml7\" (UID: \"be550258-23b4-48ba-b6e3-9d92a54b582f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417325-p9ml7" Dec 06 16:45:00 crc kubenswrapper[4813]: I1206 16:45:00.350433 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mz6j\" (UniqueName: \"kubernetes.io/projected/be550258-23b4-48ba-b6e3-9d92a54b582f-kube-api-access-8mz6j\") pod \"collect-profiles-29417325-p9ml7\" (UID: \"be550258-23b4-48ba-b6e3-9d92a54b582f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417325-p9ml7" Dec 06 16:45:00 crc kubenswrapper[4813]: I1206 16:45:00.352628 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/be550258-23b4-48ba-b6e3-9d92a54b582f-config-volume\") pod \"collect-profiles-29417325-p9ml7\" (UID: \"be550258-23b4-48ba-b6e3-9d92a54b582f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417325-p9ml7" Dec 06 16:45:00 crc kubenswrapper[4813]: I1206 16:45:00.375880 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/be550258-23b4-48ba-b6e3-9d92a54b582f-secret-volume\") pod \"collect-profiles-29417325-p9ml7\" (UID: \"be550258-23b4-48ba-b6e3-9d92a54b582f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417325-p9ml7" Dec 06 16:45:00 crc kubenswrapper[4813]: I1206 16:45:00.395792 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mz6j\" (UniqueName: \"kubernetes.io/projected/be550258-23b4-48ba-b6e3-9d92a54b582f-kube-api-access-8mz6j\") pod \"collect-profiles-29417325-p9ml7\" (UID: \"be550258-23b4-48ba-b6e3-9d92a54b582f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417325-p9ml7" Dec 06 16:45:00 crc kubenswrapper[4813]: I1206 16:45:00.546678 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 16:45:00 crc kubenswrapper[4813]: I1206 16:45:00.555191 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417325-p9ml7" Dec 06 16:45:01 crc kubenswrapper[4813]: I1206 16:45:01.058727 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417325-p9ml7"] Dec 06 16:45:01 crc kubenswrapper[4813]: W1206 16:45:01.074700 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe550258_23b4_48ba_b6e3_9d92a54b582f.slice/crio-ede531fe84aa45b622f9fd580bf4f24384f235ca9e962f34f23aa055304c8a51 WatchSource:0}: Error finding container ede531fe84aa45b622f9fd580bf4f24384f235ca9e962f34f23aa055304c8a51: Status 404 returned error can't find the container with id ede531fe84aa45b622f9fd580bf4f24384f235ca9e962f34f23aa055304c8a51 Dec 06 16:45:01 crc kubenswrapper[4813]: I1206 16:45:01.495994 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417325-p9ml7" event={"ID":"be550258-23b4-48ba-b6e3-9d92a54b582f","Type":"ContainerStarted","Data":"a57218c8b0e46f78128eb3f197354f924f2d72a7926dab9e68f4980f7d96b419"} Dec 06 16:45:01 crc kubenswrapper[4813]: I1206 16:45:01.496468 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417325-p9ml7" event={"ID":"be550258-23b4-48ba-b6e3-9d92a54b582f","Type":"ContainerStarted","Data":"ede531fe84aa45b622f9fd580bf4f24384f235ca9e962f34f23aa055304c8a51"} Dec 06 16:45:01 crc kubenswrapper[4813]: I1206 16:45:01.522078 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29417325-p9ml7" podStartSLOduration=1.5220567699999998 podStartE2EDuration="1.52205677s" podCreationTimestamp="2025-12-06 16:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:45:01.515400195 +0000 UTC m=+3541.406279781" watchObservedRunningTime="2025-12-06 16:45:01.52205677 +0000 UTC m=+3541.412936356" Dec 06 16:45:02 crc kubenswrapper[4813]: I1206 16:45:02.505618 4813 generic.go:334] "Generic (PLEG): container finished" podID="be550258-23b4-48ba-b6e3-9d92a54b582f" containerID="a57218c8b0e46f78128eb3f197354f924f2d72a7926dab9e68f4980f7d96b419" exitCode=0 Dec 06 16:45:02 crc kubenswrapper[4813]: I1206 16:45:02.505855 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417325-p9ml7" event={"ID":"be550258-23b4-48ba-b6e3-9d92a54b582f","Type":"ContainerDied","Data":"a57218c8b0e46f78128eb3f197354f924f2d72a7926dab9e68f4980f7d96b419"} Dec 06 16:45:03 crc kubenswrapper[4813]: I1206 16:45:03.906094 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417325-p9ml7" Dec 06 16:45:03 crc kubenswrapper[4813]: I1206 16:45:03.925833 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mz6j\" (UniqueName: \"kubernetes.io/projected/be550258-23b4-48ba-b6e3-9d92a54b582f-kube-api-access-8mz6j\") pod \"be550258-23b4-48ba-b6e3-9d92a54b582f\" (UID: \"be550258-23b4-48ba-b6e3-9d92a54b582f\") " Dec 06 16:45:03 crc kubenswrapper[4813]: I1206 16:45:03.925922 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/be550258-23b4-48ba-b6e3-9d92a54b582f-config-volume\") pod \"be550258-23b4-48ba-b6e3-9d92a54b582f\" (UID: \"be550258-23b4-48ba-b6e3-9d92a54b582f\") " Dec 06 16:45:03 crc kubenswrapper[4813]: I1206 16:45:03.926033 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/be550258-23b4-48ba-b6e3-9d92a54b582f-secret-volume\") pod \"be550258-23b4-48ba-b6e3-9d92a54b582f\" (UID: \"be550258-23b4-48ba-b6e3-9d92a54b582f\") " Dec 06 16:45:03 crc kubenswrapper[4813]: I1206 16:45:03.928016 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be550258-23b4-48ba-b6e3-9d92a54b582f-config-volume" (OuterVolumeSpecName: "config-volume") pod "be550258-23b4-48ba-b6e3-9d92a54b582f" (UID: "be550258-23b4-48ba-b6e3-9d92a54b582f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 16:45:03 crc kubenswrapper[4813]: I1206 16:45:03.931984 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be550258-23b4-48ba-b6e3-9d92a54b582f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "be550258-23b4-48ba-b6e3-9d92a54b582f" (UID: "be550258-23b4-48ba-b6e3-9d92a54b582f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 16:45:03 crc kubenswrapper[4813]: I1206 16:45:03.937575 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be550258-23b4-48ba-b6e3-9d92a54b582f-kube-api-access-8mz6j" (OuterVolumeSpecName: "kube-api-access-8mz6j") pod "be550258-23b4-48ba-b6e3-9d92a54b582f" (UID: "be550258-23b4-48ba-b6e3-9d92a54b582f"). InnerVolumeSpecName "kube-api-access-8mz6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:45:04 crc kubenswrapper[4813]: I1206 16:45:04.028443 4813 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/be550258-23b4-48ba-b6e3-9d92a54b582f-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 16:45:04 crc kubenswrapper[4813]: I1206 16:45:04.028685 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mz6j\" (UniqueName: \"kubernetes.io/projected/be550258-23b4-48ba-b6e3-9d92a54b582f-kube-api-access-8mz6j\") on node \"crc\" DevicePath \"\"" Dec 06 16:45:04 crc kubenswrapper[4813]: I1206 16:45:04.028758 4813 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/be550258-23b4-48ba-b6e3-9d92a54b582f-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 16:45:04 crc kubenswrapper[4813]: I1206 16:45:04.564068 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417325-p9ml7" event={"ID":"be550258-23b4-48ba-b6e3-9d92a54b582f","Type":"ContainerDied","Data":"ede531fe84aa45b622f9fd580bf4f24384f235ca9e962f34f23aa055304c8a51"} Dec 06 16:45:04 crc kubenswrapper[4813]: I1206 16:45:04.564373 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ede531fe84aa45b622f9fd580bf4f24384f235ca9e962f34f23aa055304c8a51" Dec 06 16:45:04 crc kubenswrapper[4813]: I1206 16:45:04.564454 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417325-p9ml7" Dec 06 16:45:04 crc kubenswrapper[4813]: I1206 16:45:04.601507 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417280-bd6bw"] Dec 06 16:45:04 crc kubenswrapper[4813]: I1206 16:45:04.609834 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417280-bd6bw"] Dec 06 16:45:06 crc kubenswrapper[4813]: I1206 16:45:06.510952 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="971a4ad6-5f62-47bc-b836-b7e6faaae668" path="/var/lib/kubelet/pods/971a4ad6-5f62-47bc-b836-b7e6faaae668/volumes" Dec 06 16:45:10 crc kubenswrapper[4813]: I1206 16:45:10.639721 4813 generic.go:334] "Generic (PLEG): container finished" podID="3c20161a-c3b7-4359-882f-7764fc0908ca" containerID="8152527cee71164fbc9fb5ff72a2ac2f96023937781fcb9951718f1ff107942f" exitCode=0 Dec 06 16:45:10 crc kubenswrapper[4813]: I1206 16:45:10.639818 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jlgf8/must-gather-jzpxn" event={"ID":"3c20161a-c3b7-4359-882f-7764fc0908ca","Type":"ContainerDied","Data":"8152527cee71164fbc9fb5ff72a2ac2f96023937781fcb9951718f1ff107942f"} Dec 06 16:45:10 crc kubenswrapper[4813]: I1206 16:45:10.640942 4813 scope.go:117] "RemoveContainer" containerID="8152527cee71164fbc9fb5ff72a2ac2f96023937781fcb9951718f1ff107942f" Dec 06 16:45:11 crc kubenswrapper[4813]: I1206 16:45:11.160324 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jlgf8_must-gather-jzpxn_3c20161a-c3b7-4359-882f-7764fc0908ca/gather/0.log" Dec 06 16:45:19 crc kubenswrapper[4813]: I1206 16:45:19.539473 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jlgf8/must-gather-jzpxn"] Dec 06 16:45:19 crc kubenswrapper[4813]: I1206 16:45:19.540138 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-jlgf8/must-gather-jzpxn" podUID="3c20161a-c3b7-4359-882f-7764fc0908ca" containerName="copy" containerID="cri-o://99b49e927f9e2d1e28a20400fe90995d8a0239783de2d407ed4c2825fbb52cf4" gracePeriod=2 Dec 06 16:45:19 crc kubenswrapper[4813]: I1206 16:45:19.556666 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jlgf8/must-gather-jzpxn"] Dec 06 16:45:19 crc kubenswrapper[4813]: I1206 16:45:19.732493 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jlgf8_must-gather-jzpxn_3c20161a-c3b7-4359-882f-7764fc0908ca/copy/0.log" Dec 06 16:45:19 crc kubenswrapper[4813]: I1206 16:45:19.733000 4813 generic.go:334] "Generic (PLEG): container finished" podID="3c20161a-c3b7-4359-882f-7764fc0908ca" containerID="99b49e927f9e2d1e28a20400fe90995d8a0239783de2d407ed4c2825fbb52cf4" exitCode=143 Dec 06 16:45:20 crc kubenswrapper[4813]: I1206 16:45:20.199493 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jlgf8_must-gather-jzpxn_3c20161a-c3b7-4359-882f-7764fc0908ca/copy/0.log" Dec 06 16:45:20 crc kubenswrapper[4813]: I1206 16:45:20.202680 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jlgf8/must-gather-jzpxn" Dec 06 16:45:20 crc kubenswrapper[4813]: I1206 16:45:20.393438 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cj8w9\" (UniqueName: \"kubernetes.io/projected/3c20161a-c3b7-4359-882f-7764fc0908ca-kube-api-access-cj8w9\") pod \"3c20161a-c3b7-4359-882f-7764fc0908ca\" (UID: \"3c20161a-c3b7-4359-882f-7764fc0908ca\") " Dec 06 16:45:20 crc kubenswrapper[4813]: I1206 16:45:20.393622 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3c20161a-c3b7-4359-882f-7764fc0908ca-must-gather-output\") pod \"3c20161a-c3b7-4359-882f-7764fc0908ca\" (UID: \"3c20161a-c3b7-4359-882f-7764fc0908ca\") " Dec 06 16:45:20 crc kubenswrapper[4813]: I1206 16:45:20.406875 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c20161a-c3b7-4359-882f-7764fc0908ca-kube-api-access-cj8w9" (OuterVolumeSpecName: "kube-api-access-cj8w9") pod "3c20161a-c3b7-4359-882f-7764fc0908ca" (UID: "3c20161a-c3b7-4359-882f-7764fc0908ca"). InnerVolumeSpecName "kube-api-access-cj8w9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:45:20 crc kubenswrapper[4813]: I1206 16:45:20.494900 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cj8w9\" (UniqueName: \"kubernetes.io/projected/3c20161a-c3b7-4359-882f-7764fc0908ca-kube-api-access-cj8w9\") on node \"crc\" DevicePath \"\"" Dec 06 16:45:20 crc kubenswrapper[4813]: I1206 16:45:20.533185 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c20161a-c3b7-4359-882f-7764fc0908ca-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "3c20161a-c3b7-4359-882f-7764fc0908ca" (UID: "3c20161a-c3b7-4359-882f-7764fc0908ca"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:45:20 crc kubenswrapper[4813]: I1206 16:45:20.596496 4813 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3c20161a-c3b7-4359-882f-7764fc0908ca-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 06 16:45:20 crc kubenswrapper[4813]: I1206 16:45:20.742775 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jlgf8_must-gather-jzpxn_3c20161a-c3b7-4359-882f-7764fc0908ca/copy/0.log" Dec 06 16:45:20 crc kubenswrapper[4813]: I1206 16:45:20.743300 4813 scope.go:117] "RemoveContainer" containerID="99b49e927f9e2d1e28a20400fe90995d8a0239783de2d407ed4c2825fbb52cf4" Dec 06 16:45:20 crc kubenswrapper[4813]: I1206 16:45:20.743325 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jlgf8/must-gather-jzpxn" Dec 06 16:45:20 crc kubenswrapper[4813]: I1206 16:45:20.764808 4813 scope.go:117] "RemoveContainer" containerID="8152527cee71164fbc9fb5ff72a2ac2f96023937781fcb9951718f1ff107942f" Dec 06 16:45:22 crc kubenswrapper[4813]: I1206 16:45:22.508129 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c20161a-c3b7-4359-882f-7764fc0908ca" path="/var/lib/kubelet/pods/3c20161a-c3b7-4359-882f-7764fc0908ca/volumes" Dec 06 16:45:28 crc kubenswrapper[4813]: I1206 16:45:28.379646 4813 scope.go:117] "RemoveContainer" containerID="e3f7658cda2aaf2a4a9df6e1927b6cf9c9670b5c65c1ad0d800a44fecf84f769" Dec 06 16:45:49 crc kubenswrapper[4813]: I1206 16:45:49.428032 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:45:49 crc kubenswrapper[4813]: I1206 16:45:49.428669 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:46:19 crc kubenswrapper[4813]: I1206 16:46:19.427855 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:46:19 crc kubenswrapper[4813]: I1206 16:46:19.428536 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:46:49 crc kubenswrapper[4813]: I1206 16:46:49.427805 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:46:49 crc kubenswrapper[4813]: I1206 16:46:49.430159 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:46:49 crc kubenswrapper[4813]: I1206 16:46:49.430491 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" Dec 06 16:46:49 crc kubenswrapper[4813]: I1206 16:46:49.431648 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"969c7429b9faa2222909b4bad4ef827351b31f2cde6d6ab68e5af156f1ab0dfc"} pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 16:46:49 crc kubenswrapper[4813]: I1206 16:46:49.431920 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" containerID="cri-o://969c7429b9faa2222909b4bad4ef827351b31f2cde6d6ab68e5af156f1ab0dfc" gracePeriod=600 Dec 06 16:46:49 crc kubenswrapper[4813]: E1206 16:46:49.562405 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:46:49 crc kubenswrapper[4813]: I1206 16:46:49.742965 4813 generic.go:334] "Generic (PLEG): container finished" podID="d88e8bae-c055-4c55-b548-f621ff96de06" containerID="969c7429b9faa2222909b4bad4ef827351b31f2cde6d6ab68e5af156f1ab0dfc" exitCode=0 Dec 06 16:46:49 crc kubenswrapper[4813]: I1206 16:46:49.744363 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerDied","Data":"969c7429b9faa2222909b4bad4ef827351b31f2cde6d6ab68e5af156f1ab0dfc"} Dec 06 16:46:49 crc kubenswrapper[4813]: I1206 16:46:49.744421 4813 scope.go:117] "RemoveContainer" containerID="89167fab7a057df4dbc2e1e5aa930f5ed89747de01e89ff0bb679d87ca1d7a24" Dec 06 16:46:49 crc kubenswrapper[4813]: I1206 16:46:49.744992 4813 scope.go:117] "RemoveContainer" containerID="969c7429b9faa2222909b4bad4ef827351b31f2cde6d6ab68e5af156f1ab0dfc" Dec 06 16:46:49 crc kubenswrapper[4813]: E1206 16:46:49.745211 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:47:00 crc kubenswrapper[4813]: I1206 16:47:00.490322 4813 scope.go:117] "RemoveContainer" containerID="969c7429b9faa2222909b4bad4ef827351b31f2cde6d6ab68e5af156f1ab0dfc" Dec 06 16:47:00 crc kubenswrapper[4813]: E1206 16:47:00.491099 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:47:11 crc kubenswrapper[4813]: I1206 16:47:11.486865 4813 scope.go:117] "RemoveContainer" containerID="969c7429b9faa2222909b4bad4ef827351b31f2cde6d6ab68e5af156f1ab0dfc" Dec 06 16:47:11 crc kubenswrapper[4813]: E1206 16:47:11.487679 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:47:18 crc kubenswrapper[4813]: I1206 16:47:18.248678 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rs6fc"] Dec 06 16:47:18 crc kubenswrapper[4813]: E1206 16:47:18.249581 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c20161a-c3b7-4359-882f-7764fc0908ca" containerName="copy" Dec 06 16:47:18 crc kubenswrapper[4813]: I1206 16:47:18.249592 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c20161a-c3b7-4359-882f-7764fc0908ca" containerName="copy" Dec 06 16:47:18 crc kubenswrapper[4813]: E1206 16:47:18.249604 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be550258-23b4-48ba-b6e3-9d92a54b582f" containerName="collect-profiles" Dec 06 16:47:18 crc kubenswrapper[4813]: I1206 16:47:18.249610 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="be550258-23b4-48ba-b6e3-9d92a54b582f" containerName="collect-profiles" Dec 06 16:47:18 crc kubenswrapper[4813]: E1206 16:47:18.249635 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c20161a-c3b7-4359-882f-7764fc0908ca" containerName="gather" Dec 06 16:47:18 crc kubenswrapper[4813]: I1206 16:47:18.249641 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c20161a-c3b7-4359-882f-7764fc0908ca" containerName="gather" Dec 06 16:47:18 crc kubenswrapper[4813]: I1206 16:47:18.249809 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="be550258-23b4-48ba-b6e3-9d92a54b582f" containerName="collect-profiles" Dec 06 16:47:18 crc kubenswrapper[4813]: I1206 16:47:18.249825 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c20161a-c3b7-4359-882f-7764fc0908ca" containerName="gather" Dec 06 16:47:18 crc kubenswrapper[4813]: I1206 16:47:18.249840 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c20161a-c3b7-4359-882f-7764fc0908ca" containerName="copy" Dec 06 16:47:18 crc kubenswrapper[4813]: I1206 16:47:18.251153 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rs6fc" Dec 06 16:47:18 crc kubenswrapper[4813]: I1206 16:47:18.291198 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rs6fc"] Dec 06 16:47:18 crc kubenswrapper[4813]: I1206 16:47:18.403830 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20efeeb5-a660-4753-bdce-31859a62ab2f-catalog-content\") pod \"redhat-operators-rs6fc\" (UID: \"20efeeb5-a660-4753-bdce-31859a62ab2f\") " pod="openshift-marketplace/redhat-operators-rs6fc" Dec 06 16:47:18 crc kubenswrapper[4813]: I1206 16:47:18.403913 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcb7s\" (UniqueName: \"kubernetes.io/projected/20efeeb5-a660-4753-bdce-31859a62ab2f-kube-api-access-fcb7s\") pod \"redhat-operators-rs6fc\" (UID: \"20efeeb5-a660-4753-bdce-31859a62ab2f\") " pod="openshift-marketplace/redhat-operators-rs6fc" Dec 06 16:47:18 crc kubenswrapper[4813]: I1206 16:47:18.404178 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20efeeb5-a660-4753-bdce-31859a62ab2f-utilities\") pod \"redhat-operators-rs6fc\" (UID: \"20efeeb5-a660-4753-bdce-31859a62ab2f\") " pod="openshift-marketplace/redhat-operators-rs6fc" Dec 06 16:47:18 crc kubenswrapper[4813]: I1206 16:47:18.505554 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20efeeb5-a660-4753-bdce-31859a62ab2f-catalog-content\") pod \"redhat-operators-rs6fc\" (UID: \"20efeeb5-a660-4753-bdce-31859a62ab2f\") " pod="openshift-marketplace/redhat-operators-rs6fc" Dec 06 16:47:18 crc kubenswrapper[4813]: I1206 16:47:18.505851 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcb7s\" (UniqueName: \"kubernetes.io/projected/20efeeb5-a660-4753-bdce-31859a62ab2f-kube-api-access-fcb7s\") pod \"redhat-operators-rs6fc\" (UID: \"20efeeb5-a660-4753-bdce-31859a62ab2f\") " pod="openshift-marketplace/redhat-operators-rs6fc" Dec 06 16:47:18 crc kubenswrapper[4813]: I1206 16:47:18.506052 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20efeeb5-a660-4753-bdce-31859a62ab2f-utilities\") pod \"redhat-operators-rs6fc\" (UID: \"20efeeb5-a660-4753-bdce-31859a62ab2f\") " pod="openshift-marketplace/redhat-operators-rs6fc" Dec 06 16:47:18 crc kubenswrapper[4813]: I1206 16:47:18.506065 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20efeeb5-a660-4753-bdce-31859a62ab2f-catalog-content\") pod \"redhat-operators-rs6fc\" (UID: \"20efeeb5-a660-4753-bdce-31859a62ab2f\") " pod="openshift-marketplace/redhat-operators-rs6fc" Dec 06 16:47:18 crc kubenswrapper[4813]: I1206 16:47:18.506292 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20efeeb5-a660-4753-bdce-31859a62ab2f-utilities\") pod \"redhat-operators-rs6fc\" (UID: \"20efeeb5-a660-4753-bdce-31859a62ab2f\") " pod="openshift-marketplace/redhat-operators-rs6fc" Dec 06 16:47:18 crc kubenswrapper[4813]: I1206 16:47:18.525977 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcb7s\" (UniqueName: \"kubernetes.io/projected/20efeeb5-a660-4753-bdce-31859a62ab2f-kube-api-access-fcb7s\") pod \"redhat-operators-rs6fc\" (UID: \"20efeeb5-a660-4753-bdce-31859a62ab2f\") " pod="openshift-marketplace/redhat-operators-rs6fc" Dec 06 16:47:18 crc kubenswrapper[4813]: I1206 16:47:18.576299 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rs6fc" Dec 06 16:47:19 crc kubenswrapper[4813]: I1206 16:47:19.051057 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rs6fc"] Dec 06 16:47:20 crc kubenswrapper[4813]: I1206 16:47:20.035869 4813 generic.go:334] "Generic (PLEG): container finished" podID="20efeeb5-a660-4753-bdce-31859a62ab2f" containerID="1cd987b149ae3ff4a1fc7b1d332637f4e4cf2ff991af1f7d31e7bab040919be6" exitCode=0 Dec 06 16:47:20 crc kubenswrapper[4813]: I1206 16:47:20.036182 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rs6fc" event={"ID":"20efeeb5-a660-4753-bdce-31859a62ab2f","Type":"ContainerDied","Data":"1cd987b149ae3ff4a1fc7b1d332637f4e4cf2ff991af1f7d31e7bab040919be6"} Dec 06 16:47:20 crc kubenswrapper[4813]: I1206 16:47:20.036227 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rs6fc" event={"ID":"20efeeb5-a660-4753-bdce-31859a62ab2f","Type":"ContainerStarted","Data":"6575b0f7663ef5a1f44517d475427216b44e70dfda3d51c1115ba907445a214d"} Dec 06 16:47:21 crc kubenswrapper[4813]: I1206 16:47:21.047976 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rs6fc" event={"ID":"20efeeb5-a660-4753-bdce-31859a62ab2f","Type":"ContainerStarted","Data":"3870bb9bedfefb4aae2b642d25365366d73faf467397d1ff13c7c9b9c95b31d7"} Dec 06 16:47:24 crc kubenswrapper[4813]: I1206 16:47:24.091576 4813 generic.go:334] "Generic (PLEG): container finished" podID="20efeeb5-a660-4753-bdce-31859a62ab2f" containerID="3870bb9bedfefb4aae2b642d25365366d73faf467397d1ff13c7c9b9c95b31d7" exitCode=0 Dec 06 16:47:24 crc kubenswrapper[4813]: I1206 16:47:24.091633 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rs6fc" event={"ID":"20efeeb5-a660-4753-bdce-31859a62ab2f","Type":"ContainerDied","Data":"3870bb9bedfefb4aae2b642d25365366d73faf467397d1ff13c7c9b9c95b31d7"} Dec 06 16:47:25 crc kubenswrapper[4813]: I1206 16:47:25.103224 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rs6fc" event={"ID":"20efeeb5-a660-4753-bdce-31859a62ab2f","Type":"ContainerStarted","Data":"23a8bf6116d31cb77e898d647668da9e93fcf6166d8dae8fd6adf1cf486fa448"} Dec 06 16:47:25 crc kubenswrapper[4813]: I1206 16:47:25.128247 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rs6fc" podStartSLOduration=2.581561208 podStartE2EDuration="7.128225445s" podCreationTimestamp="2025-12-06 16:47:18 +0000 UTC" firstStartedPulling="2025-12-06 16:47:20.039909274 +0000 UTC m=+3679.930788870" lastFinishedPulling="2025-12-06 16:47:24.586573501 +0000 UTC m=+3684.477453107" observedRunningTime="2025-12-06 16:47:25.124045425 +0000 UTC m=+3685.014925031" watchObservedRunningTime="2025-12-06 16:47:25.128225445 +0000 UTC m=+3685.019105051" Dec 06 16:47:26 crc kubenswrapper[4813]: I1206 16:47:26.487902 4813 scope.go:117] "RemoveContainer" containerID="969c7429b9faa2222909b4bad4ef827351b31f2cde6d6ab68e5af156f1ab0dfc" Dec 06 16:47:26 crc kubenswrapper[4813]: E1206 16:47:26.488757 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:47:28 crc kubenswrapper[4813]: I1206 16:47:28.576896 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rs6fc" Dec 06 16:47:28 crc kubenswrapper[4813]: I1206 16:47:28.577190 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rs6fc" Dec 06 16:47:29 crc kubenswrapper[4813]: I1206 16:47:29.648402 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rs6fc" podUID="20efeeb5-a660-4753-bdce-31859a62ab2f" containerName="registry-server" probeResult="failure" output=< Dec 06 16:47:29 crc kubenswrapper[4813]: timeout: failed to connect service ":50051" within 1s Dec 06 16:47:29 crc kubenswrapper[4813]: > Dec 06 16:47:37 crc kubenswrapper[4813]: I1206 16:47:37.486950 4813 scope.go:117] "RemoveContainer" containerID="969c7429b9faa2222909b4bad4ef827351b31f2cde6d6ab68e5af156f1ab0dfc" Dec 06 16:47:37 crc kubenswrapper[4813]: E1206 16:47:37.487733 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:47:38 crc kubenswrapper[4813]: I1206 16:47:38.661336 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rs6fc" Dec 06 16:47:38 crc kubenswrapper[4813]: I1206 16:47:38.749820 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rs6fc" Dec 06 16:47:38 crc kubenswrapper[4813]: I1206 16:47:38.917648 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rs6fc"] Dec 06 16:47:40 crc kubenswrapper[4813]: I1206 16:47:40.273552 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rs6fc" podUID="20efeeb5-a660-4753-bdce-31859a62ab2f" containerName="registry-server" containerID="cri-o://23a8bf6116d31cb77e898d647668da9e93fcf6166d8dae8fd6adf1cf486fa448" gracePeriod=2 Dec 06 16:47:40 crc kubenswrapper[4813]: I1206 16:47:40.746600 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rs6fc" Dec 06 16:47:40 crc kubenswrapper[4813]: I1206 16:47:40.872840 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20efeeb5-a660-4753-bdce-31859a62ab2f-catalog-content\") pod \"20efeeb5-a660-4753-bdce-31859a62ab2f\" (UID: \"20efeeb5-a660-4753-bdce-31859a62ab2f\") " Dec 06 16:47:40 crc kubenswrapper[4813]: I1206 16:47:40.873129 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20efeeb5-a660-4753-bdce-31859a62ab2f-utilities\") pod \"20efeeb5-a660-4753-bdce-31859a62ab2f\" (UID: \"20efeeb5-a660-4753-bdce-31859a62ab2f\") " Dec 06 16:47:40 crc kubenswrapper[4813]: I1206 16:47:40.873177 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcb7s\" (UniqueName: \"kubernetes.io/projected/20efeeb5-a660-4753-bdce-31859a62ab2f-kube-api-access-fcb7s\") pod \"20efeeb5-a660-4753-bdce-31859a62ab2f\" (UID: \"20efeeb5-a660-4753-bdce-31859a62ab2f\") " Dec 06 16:47:40 crc kubenswrapper[4813]: I1206 16:47:40.874479 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20efeeb5-a660-4753-bdce-31859a62ab2f-utilities" (OuterVolumeSpecName: "utilities") pod "20efeeb5-a660-4753-bdce-31859a62ab2f" (UID: "20efeeb5-a660-4753-bdce-31859a62ab2f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:47:40 crc kubenswrapper[4813]: I1206 16:47:40.881475 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20efeeb5-a660-4753-bdce-31859a62ab2f-kube-api-access-fcb7s" (OuterVolumeSpecName: "kube-api-access-fcb7s") pod "20efeeb5-a660-4753-bdce-31859a62ab2f" (UID: "20efeeb5-a660-4753-bdce-31859a62ab2f"). InnerVolumeSpecName "kube-api-access-fcb7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:47:40 crc kubenswrapper[4813]: I1206 16:47:40.975778 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20efeeb5-a660-4753-bdce-31859a62ab2f-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 16:47:40 crc kubenswrapper[4813]: I1206 16:47:40.975846 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcb7s\" (UniqueName: \"kubernetes.io/projected/20efeeb5-a660-4753-bdce-31859a62ab2f-kube-api-access-fcb7s\") on node \"crc\" DevicePath \"\"" Dec 06 16:47:40 crc kubenswrapper[4813]: I1206 16:47:40.990832 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20efeeb5-a660-4753-bdce-31859a62ab2f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "20efeeb5-a660-4753-bdce-31859a62ab2f" (UID: "20efeeb5-a660-4753-bdce-31859a62ab2f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:47:41 crc kubenswrapper[4813]: I1206 16:47:41.077254 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20efeeb5-a660-4753-bdce-31859a62ab2f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 16:47:41 crc kubenswrapper[4813]: I1206 16:47:41.290003 4813 generic.go:334] "Generic (PLEG): container finished" podID="20efeeb5-a660-4753-bdce-31859a62ab2f" containerID="23a8bf6116d31cb77e898d647668da9e93fcf6166d8dae8fd6adf1cf486fa448" exitCode=0 Dec 06 16:47:41 crc kubenswrapper[4813]: I1206 16:47:41.290078 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rs6fc" event={"ID":"20efeeb5-a660-4753-bdce-31859a62ab2f","Type":"ContainerDied","Data":"23a8bf6116d31cb77e898d647668da9e93fcf6166d8dae8fd6adf1cf486fa448"} Dec 06 16:47:41 crc kubenswrapper[4813]: I1206 16:47:41.290126 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rs6fc" event={"ID":"20efeeb5-a660-4753-bdce-31859a62ab2f","Type":"ContainerDied","Data":"6575b0f7663ef5a1f44517d475427216b44e70dfda3d51c1115ba907445a214d"} Dec 06 16:47:41 crc kubenswrapper[4813]: I1206 16:47:41.290143 4813 scope.go:117] "RemoveContainer" containerID="23a8bf6116d31cb77e898d647668da9e93fcf6166d8dae8fd6adf1cf486fa448" Dec 06 16:47:41 crc kubenswrapper[4813]: I1206 16:47:41.290141 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rs6fc" Dec 06 16:47:41 crc kubenswrapper[4813]: I1206 16:47:41.322280 4813 scope.go:117] "RemoveContainer" containerID="3870bb9bedfefb4aae2b642d25365366d73faf467397d1ff13c7c9b9c95b31d7" Dec 06 16:47:41 crc kubenswrapper[4813]: I1206 16:47:41.345981 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rs6fc"] Dec 06 16:47:41 crc kubenswrapper[4813]: I1206 16:47:41.351796 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rs6fc"] Dec 06 16:47:41 crc kubenswrapper[4813]: I1206 16:47:41.378948 4813 scope.go:117] "RemoveContainer" containerID="1cd987b149ae3ff4a1fc7b1d332637f4e4cf2ff991af1f7d31e7bab040919be6" Dec 06 16:47:41 crc kubenswrapper[4813]: I1206 16:47:41.422745 4813 scope.go:117] "RemoveContainer" containerID="23a8bf6116d31cb77e898d647668da9e93fcf6166d8dae8fd6adf1cf486fa448" Dec 06 16:47:41 crc kubenswrapper[4813]: E1206 16:47:41.424742 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23a8bf6116d31cb77e898d647668da9e93fcf6166d8dae8fd6adf1cf486fa448\": container with ID starting with 23a8bf6116d31cb77e898d647668da9e93fcf6166d8dae8fd6adf1cf486fa448 not found: ID does not exist" containerID="23a8bf6116d31cb77e898d647668da9e93fcf6166d8dae8fd6adf1cf486fa448" Dec 06 16:47:41 crc kubenswrapper[4813]: I1206 16:47:41.424785 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23a8bf6116d31cb77e898d647668da9e93fcf6166d8dae8fd6adf1cf486fa448"} err="failed to get container status \"23a8bf6116d31cb77e898d647668da9e93fcf6166d8dae8fd6adf1cf486fa448\": rpc error: code = NotFound desc = could not find container \"23a8bf6116d31cb77e898d647668da9e93fcf6166d8dae8fd6adf1cf486fa448\": container with ID starting with 23a8bf6116d31cb77e898d647668da9e93fcf6166d8dae8fd6adf1cf486fa448 not found: ID does not exist" Dec 06 16:47:41 crc kubenswrapper[4813]: I1206 16:47:41.424813 4813 scope.go:117] "RemoveContainer" containerID="3870bb9bedfefb4aae2b642d25365366d73faf467397d1ff13c7c9b9c95b31d7" Dec 06 16:47:41 crc kubenswrapper[4813]: E1206 16:47:41.425309 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3870bb9bedfefb4aae2b642d25365366d73faf467397d1ff13c7c9b9c95b31d7\": container with ID starting with 3870bb9bedfefb4aae2b642d25365366d73faf467397d1ff13c7c9b9c95b31d7 not found: ID does not exist" containerID="3870bb9bedfefb4aae2b642d25365366d73faf467397d1ff13c7c9b9c95b31d7" Dec 06 16:47:41 crc kubenswrapper[4813]: I1206 16:47:41.425360 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3870bb9bedfefb4aae2b642d25365366d73faf467397d1ff13c7c9b9c95b31d7"} err="failed to get container status \"3870bb9bedfefb4aae2b642d25365366d73faf467397d1ff13c7c9b9c95b31d7\": rpc error: code = NotFound desc = could not find container \"3870bb9bedfefb4aae2b642d25365366d73faf467397d1ff13c7c9b9c95b31d7\": container with ID starting with 3870bb9bedfefb4aae2b642d25365366d73faf467397d1ff13c7c9b9c95b31d7 not found: ID does not exist" Dec 06 16:47:41 crc kubenswrapper[4813]: I1206 16:47:41.425392 4813 scope.go:117] "RemoveContainer" containerID="1cd987b149ae3ff4a1fc7b1d332637f4e4cf2ff991af1f7d31e7bab040919be6" Dec 06 16:47:41 crc kubenswrapper[4813]: E1206 16:47:41.425728 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1cd987b149ae3ff4a1fc7b1d332637f4e4cf2ff991af1f7d31e7bab040919be6\": container with ID starting with 1cd987b149ae3ff4a1fc7b1d332637f4e4cf2ff991af1f7d31e7bab040919be6 not found: ID does not exist" containerID="1cd987b149ae3ff4a1fc7b1d332637f4e4cf2ff991af1f7d31e7bab040919be6" Dec 06 16:47:41 crc kubenswrapper[4813]: I1206 16:47:41.425757 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cd987b149ae3ff4a1fc7b1d332637f4e4cf2ff991af1f7d31e7bab040919be6"} err="failed to get container status \"1cd987b149ae3ff4a1fc7b1d332637f4e4cf2ff991af1f7d31e7bab040919be6\": rpc error: code = NotFound desc = could not find container \"1cd987b149ae3ff4a1fc7b1d332637f4e4cf2ff991af1f7d31e7bab040919be6\": container with ID starting with 1cd987b149ae3ff4a1fc7b1d332637f4e4cf2ff991af1f7d31e7bab040919be6 not found: ID does not exist" Dec 06 16:47:42 crc kubenswrapper[4813]: I1206 16:47:42.502473 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20efeeb5-a660-4753-bdce-31859a62ab2f" path="/var/lib/kubelet/pods/20efeeb5-a660-4753-bdce-31859a62ab2f/volumes" Dec 06 16:47:52 crc kubenswrapper[4813]: I1206 16:47:52.491239 4813 scope.go:117] "RemoveContainer" containerID="969c7429b9faa2222909b4bad4ef827351b31f2cde6d6ab68e5af156f1ab0dfc" Dec 06 16:47:52 crc kubenswrapper[4813]: E1206 16:47:52.492336 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:48:04 crc kubenswrapper[4813]: I1206 16:48:04.487756 4813 scope.go:117] "RemoveContainer" containerID="969c7429b9faa2222909b4bad4ef827351b31f2cde6d6ab68e5af156f1ab0dfc" Dec 06 16:48:04 crc kubenswrapper[4813]: E1206 16:48:04.490615 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:48:15 crc kubenswrapper[4813]: I1206 16:48:15.487908 4813 scope.go:117] "RemoveContainer" containerID="969c7429b9faa2222909b4bad4ef827351b31f2cde6d6ab68e5af156f1ab0dfc" Dec 06 16:48:15 crc kubenswrapper[4813]: E1206 16:48:15.488608 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:48:18 crc kubenswrapper[4813]: I1206 16:48:18.181002 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jgn6l/must-gather-jb2rq"] Dec 06 16:48:18 crc kubenswrapper[4813]: E1206 16:48:18.181618 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20efeeb5-a660-4753-bdce-31859a62ab2f" containerName="extract-content" Dec 06 16:48:18 crc kubenswrapper[4813]: I1206 16:48:18.181630 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="20efeeb5-a660-4753-bdce-31859a62ab2f" containerName="extract-content" Dec 06 16:48:18 crc kubenswrapper[4813]: E1206 16:48:18.181648 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20efeeb5-a660-4753-bdce-31859a62ab2f" containerName="extract-utilities" Dec 06 16:48:18 crc kubenswrapper[4813]: I1206 16:48:18.181655 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="20efeeb5-a660-4753-bdce-31859a62ab2f" containerName="extract-utilities" Dec 06 16:48:18 crc kubenswrapper[4813]: E1206 16:48:18.181677 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20efeeb5-a660-4753-bdce-31859a62ab2f" containerName="registry-server" Dec 06 16:48:18 crc kubenswrapper[4813]: I1206 16:48:18.181684 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="20efeeb5-a660-4753-bdce-31859a62ab2f" containerName="registry-server" Dec 06 16:48:18 crc kubenswrapper[4813]: I1206 16:48:18.181866 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="20efeeb5-a660-4753-bdce-31859a62ab2f" containerName="registry-server" Dec 06 16:48:18 crc kubenswrapper[4813]: I1206 16:48:18.182842 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgn6l/must-gather-jb2rq" Dec 06 16:48:18 crc kubenswrapper[4813]: I1206 16:48:18.192481 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-jgn6l/must-gather-jb2rq"] Dec 06 16:48:18 crc kubenswrapper[4813]: I1206 16:48:18.220228 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-jgn6l"/"openshift-service-ca.crt" Dec 06 16:48:18 crc kubenswrapper[4813]: I1206 16:48:18.221597 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-jgn6l"/"default-dockercfg-tpx8v" Dec 06 16:48:18 crc kubenswrapper[4813]: I1206 16:48:18.221826 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-jgn6l"/"kube-root-ca.crt" Dec 06 16:48:18 crc kubenswrapper[4813]: I1206 16:48:18.354194 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1a835870-e23a-446e-8c38-fe587e9cee6b-must-gather-output\") pod \"must-gather-jb2rq\" (UID: \"1a835870-e23a-446e-8c38-fe587e9cee6b\") " pod="openshift-must-gather-jgn6l/must-gather-jb2rq" Dec 06 16:48:18 crc kubenswrapper[4813]: I1206 16:48:18.354389 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndcqn\" (UniqueName: \"kubernetes.io/projected/1a835870-e23a-446e-8c38-fe587e9cee6b-kube-api-access-ndcqn\") pod \"must-gather-jb2rq\" (UID: \"1a835870-e23a-446e-8c38-fe587e9cee6b\") " pod="openshift-must-gather-jgn6l/must-gather-jb2rq" Dec 06 16:48:18 crc kubenswrapper[4813]: I1206 16:48:18.455961 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndcqn\" (UniqueName: \"kubernetes.io/projected/1a835870-e23a-446e-8c38-fe587e9cee6b-kube-api-access-ndcqn\") pod \"must-gather-jb2rq\" (UID: \"1a835870-e23a-446e-8c38-fe587e9cee6b\") " pod="openshift-must-gather-jgn6l/must-gather-jb2rq" Dec 06 16:48:18 crc kubenswrapper[4813]: I1206 16:48:18.456051 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1a835870-e23a-446e-8c38-fe587e9cee6b-must-gather-output\") pod \"must-gather-jb2rq\" (UID: \"1a835870-e23a-446e-8c38-fe587e9cee6b\") " pod="openshift-must-gather-jgn6l/must-gather-jb2rq" Dec 06 16:48:18 crc kubenswrapper[4813]: I1206 16:48:18.456519 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1a835870-e23a-446e-8c38-fe587e9cee6b-must-gather-output\") pod \"must-gather-jb2rq\" (UID: \"1a835870-e23a-446e-8c38-fe587e9cee6b\") " pod="openshift-must-gather-jgn6l/must-gather-jb2rq" Dec 06 16:48:18 crc kubenswrapper[4813]: I1206 16:48:18.473390 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndcqn\" (UniqueName: \"kubernetes.io/projected/1a835870-e23a-446e-8c38-fe587e9cee6b-kube-api-access-ndcqn\") pod \"must-gather-jb2rq\" (UID: \"1a835870-e23a-446e-8c38-fe587e9cee6b\") " pod="openshift-must-gather-jgn6l/must-gather-jb2rq" Dec 06 16:48:18 crc kubenswrapper[4813]: I1206 16:48:18.553035 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgn6l/must-gather-jb2rq" Dec 06 16:48:18 crc kubenswrapper[4813]: I1206 16:48:18.807003 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-jgn6l/must-gather-jb2rq"] Dec 06 16:48:19 crc kubenswrapper[4813]: I1206 16:48:19.786102 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jgn6l/must-gather-jb2rq" event={"ID":"1a835870-e23a-446e-8c38-fe587e9cee6b","Type":"ContainerStarted","Data":"e3d87520cb39ce4ee89b161ddfb351a6c242a2718c3f771de3ebdafd24aec70e"} Dec 06 16:48:19 crc kubenswrapper[4813]: I1206 16:48:19.786447 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jgn6l/must-gather-jb2rq" event={"ID":"1a835870-e23a-446e-8c38-fe587e9cee6b","Type":"ContainerStarted","Data":"db9b299109c15116a10e92800ddbe7711c0fbc10c48090abc39a6e10048f6867"} Dec 06 16:48:19 crc kubenswrapper[4813]: I1206 16:48:19.786460 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jgn6l/must-gather-jb2rq" event={"ID":"1a835870-e23a-446e-8c38-fe587e9cee6b","Type":"ContainerStarted","Data":"2002d2ee2073d38c827c66dc0d99a6da2fc04f35fd44bdd1d4b4a809d4c5d41d"} Dec 06 16:48:19 crc kubenswrapper[4813]: I1206 16:48:19.808686 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-jgn6l/must-gather-jb2rq" podStartSLOduration=1.808668663 podStartE2EDuration="1.808668663s" podCreationTimestamp="2025-12-06 16:48:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:48:19.80701964 +0000 UTC m=+3739.697899216" watchObservedRunningTime="2025-12-06 16:48:19.808668663 +0000 UTC m=+3739.699548239" Dec 06 16:48:22 crc kubenswrapper[4813]: I1206 16:48:22.741465 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jgn6l/crc-debug-dhxwz"] Dec 06 16:48:22 crc kubenswrapper[4813]: I1206 16:48:22.742922 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgn6l/crc-debug-dhxwz" Dec 06 16:48:22 crc kubenswrapper[4813]: I1206 16:48:22.831816 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9defbb7b-e914-406e-a74c-cc3cb1885471-host\") pod \"crc-debug-dhxwz\" (UID: \"9defbb7b-e914-406e-a74c-cc3cb1885471\") " pod="openshift-must-gather-jgn6l/crc-debug-dhxwz" Dec 06 16:48:22 crc kubenswrapper[4813]: I1206 16:48:22.831878 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcss8\" (UniqueName: \"kubernetes.io/projected/9defbb7b-e914-406e-a74c-cc3cb1885471-kube-api-access-gcss8\") pod \"crc-debug-dhxwz\" (UID: \"9defbb7b-e914-406e-a74c-cc3cb1885471\") " pod="openshift-must-gather-jgn6l/crc-debug-dhxwz" Dec 06 16:48:22 crc kubenswrapper[4813]: I1206 16:48:22.933371 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9defbb7b-e914-406e-a74c-cc3cb1885471-host\") pod \"crc-debug-dhxwz\" (UID: \"9defbb7b-e914-406e-a74c-cc3cb1885471\") " pod="openshift-must-gather-jgn6l/crc-debug-dhxwz" Dec 06 16:48:22 crc kubenswrapper[4813]: I1206 16:48:22.933438 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcss8\" (UniqueName: \"kubernetes.io/projected/9defbb7b-e914-406e-a74c-cc3cb1885471-kube-api-access-gcss8\") pod \"crc-debug-dhxwz\" (UID: \"9defbb7b-e914-406e-a74c-cc3cb1885471\") " pod="openshift-must-gather-jgn6l/crc-debug-dhxwz" Dec 06 16:48:22 crc kubenswrapper[4813]: I1206 16:48:22.933877 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9defbb7b-e914-406e-a74c-cc3cb1885471-host\") pod \"crc-debug-dhxwz\" (UID: \"9defbb7b-e914-406e-a74c-cc3cb1885471\") " pod="openshift-must-gather-jgn6l/crc-debug-dhxwz" Dec 06 16:48:22 crc kubenswrapper[4813]: I1206 16:48:22.956125 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcss8\" (UniqueName: \"kubernetes.io/projected/9defbb7b-e914-406e-a74c-cc3cb1885471-kube-api-access-gcss8\") pod \"crc-debug-dhxwz\" (UID: \"9defbb7b-e914-406e-a74c-cc3cb1885471\") " pod="openshift-must-gather-jgn6l/crc-debug-dhxwz" Dec 06 16:48:23 crc kubenswrapper[4813]: I1206 16:48:23.068213 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgn6l/crc-debug-dhxwz" Dec 06 16:48:23 crc kubenswrapper[4813]: W1206 16:48:23.101350 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9defbb7b_e914_406e_a74c_cc3cb1885471.slice/crio-a47349408324664a1b143a5783598e252a0e21c37b5f51139aa251f6a75044c4 WatchSource:0}: Error finding container a47349408324664a1b143a5783598e252a0e21c37b5f51139aa251f6a75044c4: Status 404 returned error can't find the container with id a47349408324664a1b143a5783598e252a0e21c37b5f51139aa251f6a75044c4 Dec 06 16:48:23 crc kubenswrapper[4813]: I1206 16:48:23.831436 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jgn6l/crc-debug-dhxwz" event={"ID":"9defbb7b-e914-406e-a74c-cc3cb1885471","Type":"ContainerStarted","Data":"298032040ae711ef3a87d5323381a6219fb27bfa2e281d7cc14b6bd2843067a5"} Dec 06 16:48:23 crc kubenswrapper[4813]: I1206 16:48:23.832648 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jgn6l/crc-debug-dhxwz" event={"ID":"9defbb7b-e914-406e-a74c-cc3cb1885471","Type":"ContainerStarted","Data":"a47349408324664a1b143a5783598e252a0e21c37b5f51139aa251f6a75044c4"} Dec 06 16:48:23 crc kubenswrapper[4813]: I1206 16:48:23.850996 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-jgn6l/crc-debug-dhxwz" podStartSLOduration=1.850977906 podStartE2EDuration="1.850977906s" podCreationTimestamp="2025-12-06 16:48:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 16:48:23.850423022 +0000 UTC m=+3743.741302598" watchObservedRunningTime="2025-12-06 16:48:23.850977906 +0000 UTC m=+3743.741857482" Dec 06 16:48:28 crc kubenswrapper[4813]: I1206 16:48:28.487054 4813 scope.go:117] "RemoveContainer" containerID="969c7429b9faa2222909b4bad4ef827351b31f2cde6d6ab68e5af156f1ab0dfc" Dec 06 16:48:28 crc kubenswrapper[4813]: E1206 16:48:28.487953 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:48:40 crc kubenswrapper[4813]: I1206 16:48:40.492171 4813 scope.go:117] "RemoveContainer" containerID="969c7429b9faa2222909b4bad4ef827351b31f2cde6d6ab68e5af156f1ab0dfc" Dec 06 16:48:40 crc kubenswrapper[4813]: E1206 16:48:40.493032 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:48:54 crc kubenswrapper[4813]: I1206 16:48:54.487576 4813 scope.go:117] "RemoveContainer" containerID="969c7429b9faa2222909b4bad4ef827351b31f2cde6d6ab68e5af156f1ab0dfc" Dec 06 16:48:54 crc kubenswrapper[4813]: E1206 16:48:54.488240 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:48:55 crc kubenswrapper[4813]: I1206 16:48:55.081153 4813 generic.go:334] "Generic (PLEG): container finished" podID="9defbb7b-e914-406e-a74c-cc3cb1885471" containerID="298032040ae711ef3a87d5323381a6219fb27bfa2e281d7cc14b6bd2843067a5" exitCode=0 Dec 06 16:48:55 crc kubenswrapper[4813]: I1206 16:48:55.081251 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jgn6l/crc-debug-dhxwz" event={"ID":"9defbb7b-e914-406e-a74c-cc3cb1885471","Type":"ContainerDied","Data":"298032040ae711ef3a87d5323381a6219fb27bfa2e281d7cc14b6bd2843067a5"} Dec 06 16:48:56 crc kubenswrapper[4813]: I1206 16:48:56.197691 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgn6l/crc-debug-dhxwz" Dec 06 16:48:56 crc kubenswrapper[4813]: I1206 16:48:56.226981 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jgn6l/crc-debug-dhxwz"] Dec 06 16:48:56 crc kubenswrapper[4813]: I1206 16:48:56.236387 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jgn6l/crc-debug-dhxwz"] Dec 06 16:48:56 crc kubenswrapper[4813]: I1206 16:48:56.391530 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcss8\" (UniqueName: \"kubernetes.io/projected/9defbb7b-e914-406e-a74c-cc3cb1885471-kube-api-access-gcss8\") pod \"9defbb7b-e914-406e-a74c-cc3cb1885471\" (UID: \"9defbb7b-e914-406e-a74c-cc3cb1885471\") " Dec 06 16:48:56 crc kubenswrapper[4813]: I1206 16:48:56.391726 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9defbb7b-e914-406e-a74c-cc3cb1885471-host\") pod \"9defbb7b-e914-406e-a74c-cc3cb1885471\" (UID: \"9defbb7b-e914-406e-a74c-cc3cb1885471\") " Dec 06 16:48:56 crc kubenswrapper[4813]: I1206 16:48:56.391809 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9defbb7b-e914-406e-a74c-cc3cb1885471-host" (OuterVolumeSpecName: "host") pod "9defbb7b-e914-406e-a74c-cc3cb1885471" (UID: "9defbb7b-e914-406e-a74c-cc3cb1885471"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 16:48:56 crc kubenswrapper[4813]: I1206 16:48:56.392387 4813 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9defbb7b-e914-406e-a74c-cc3cb1885471-host\") on node \"crc\" DevicePath \"\"" Dec 06 16:48:56 crc kubenswrapper[4813]: I1206 16:48:56.402685 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9defbb7b-e914-406e-a74c-cc3cb1885471-kube-api-access-gcss8" (OuterVolumeSpecName: "kube-api-access-gcss8") pod "9defbb7b-e914-406e-a74c-cc3cb1885471" (UID: "9defbb7b-e914-406e-a74c-cc3cb1885471"). InnerVolumeSpecName "kube-api-access-gcss8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:48:56 crc kubenswrapper[4813]: I1206 16:48:56.493409 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcss8\" (UniqueName: \"kubernetes.io/projected/9defbb7b-e914-406e-a74c-cc3cb1885471-kube-api-access-gcss8\") on node \"crc\" DevicePath \"\"" Dec 06 16:48:56 crc kubenswrapper[4813]: I1206 16:48:56.496457 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9defbb7b-e914-406e-a74c-cc3cb1885471" path="/var/lib/kubelet/pods/9defbb7b-e914-406e-a74c-cc3cb1885471/volumes" Dec 06 16:48:57 crc kubenswrapper[4813]: I1206 16:48:57.095508 4813 scope.go:117] "RemoveContainer" containerID="298032040ae711ef3a87d5323381a6219fb27bfa2e281d7cc14b6bd2843067a5" Dec 06 16:48:57 crc kubenswrapper[4813]: I1206 16:48:57.095637 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgn6l/crc-debug-dhxwz" Dec 06 16:48:57 crc kubenswrapper[4813]: I1206 16:48:57.528246 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jgn6l/crc-debug-4p6mj"] Dec 06 16:48:57 crc kubenswrapper[4813]: E1206 16:48:57.529303 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9defbb7b-e914-406e-a74c-cc3cb1885471" containerName="container-00" Dec 06 16:48:57 crc kubenswrapper[4813]: I1206 16:48:57.529381 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="9defbb7b-e914-406e-a74c-cc3cb1885471" containerName="container-00" Dec 06 16:48:57 crc kubenswrapper[4813]: I1206 16:48:57.529604 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="9defbb7b-e914-406e-a74c-cc3cb1885471" containerName="container-00" Dec 06 16:48:57 crc kubenswrapper[4813]: I1206 16:48:57.530229 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgn6l/crc-debug-4p6mj" Dec 06 16:48:57 crc kubenswrapper[4813]: I1206 16:48:57.713283 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kh6v\" (UniqueName: \"kubernetes.io/projected/18497f1a-4db2-445a-9286-8a331fe18c30-kube-api-access-8kh6v\") pod \"crc-debug-4p6mj\" (UID: \"18497f1a-4db2-445a-9286-8a331fe18c30\") " pod="openshift-must-gather-jgn6l/crc-debug-4p6mj" Dec 06 16:48:57 crc kubenswrapper[4813]: I1206 16:48:57.714022 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/18497f1a-4db2-445a-9286-8a331fe18c30-host\") pod \"crc-debug-4p6mj\" (UID: \"18497f1a-4db2-445a-9286-8a331fe18c30\") " pod="openshift-must-gather-jgn6l/crc-debug-4p6mj" Dec 06 16:48:57 crc kubenswrapper[4813]: I1206 16:48:57.815891 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kh6v\" (UniqueName: \"kubernetes.io/projected/18497f1a-4db2-445a-9286-8a331fe18c30-kube-api-access-8kh6v\") pod \"crc-debug-4p6mj\" (UID: \"18497f1a-4db2-445a-9286-8a331fe18c30\") " pod="openshift-must-gather-jgn6l/crc-debug-4p6mj" Dec 06 16:48:57 crc kubenswrapper[4813]: I1206 16:48:57.815978 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/18497f1a-4db2-445a-9286-8a331fe18c30-host\") pod \"crc-debug-4p6mj\" (UID: \"18497f1a-4db2-445a-9286-8a331fe18c30\") " pod="openshift-must-gather-jgn6l/crc-debug-4p6mj" Dec 06 16:48:57 crc kubenswrapper[4813]: I1206 16:48:57.816083 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/18497f1a-4db2-445a-9286-8a331fe18c30-host\") pod \"crc-debug-4p6mj\" (UID: \"18497f1a-4db2-445a-9286-8a331fe18c30\") " pod="openshift-must-gather-jgn6l/crc-debug-4p6mj" Dec 06 16:48:57 crc kubenswrapper[4813]: I1206 16:48:57.838030 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kh6v\" (UniqueName: \"kubernetes.io/projected/18497f1a-4db2-445a-9286-8a331fe18c30-kube-api-access-8kh6v\") pod \"crc-debug-4p6mj\" (UID: \"18497f1a-4db2-445a-9286-8a331fe18c30\") " pod="openshift-must-gather-jgn6l/crc-debug-4p6mj" Dec 06 16:48:57 crc kubenswrapper[4813]: I1206 16:48:57.842984 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgn6l/crc-debug-4p6mj" Dec 06 16:48:58 crc kubenswrapper[4813]: I1206 16:48:58.105475 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jgn6l/crc-debug-4p6mj" event={"ID":"18497f1a-4db2-445a-9286-8a331fe18c30","Type":"ContainerStarted","Data":"99f91363a0116b1813a6b42e4861f9e144ae96c33c70212ed502353488f63807"} Dec 06 16:48:58 crc kubenswrapper[4813]: I1206 16:48:58.105839 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jgn6l/crc-debug-4p6mj" event={"ID":"18497f1a-4db2-445a-9286-8a331fe18c30","Type":"ContainerStarted","Data":"4357dfa0940dbadd517df5d8fe28654a42e5e3d19cfa5bee1ed91d54c49faa4c"} Dec 06 16:48:59 crc kubenswrapper[4813]: I1206 16:48:59.116472 4813 generic.go:334] "Generic (PLEG): container finished" podID="18497f1a-4db2-445a-9286-8a331fe18c30" containerID="99f91363a0116b1813a6b42e4861f9e144ae96c33c70212ed502353488f63807" exitCode=0 Dec 06 16:48:59 crc kubenswrapper[4813]: I1206 16:48:59.116610 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jgn6l/crc-debug-4p6mj" event={"ID":"18497f1a-4db2-445a-9286-8a331fe18c30","Type":"ContainerDied","Data":"99f91363a0116b1813a6b42e4861f9e144ae96c33c70212ed502353488f63807"} Dec 06 16:49:00 crc kubenswrapper[4813]: I1206 16:49:00.210814 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgn6l/crc-debug-4p6mj" Dec 06 16:49:00 crc kubenswrapper[4813]: I1206 16:49:00.242086 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jgn6l/crc-debug-4p6mj"] Dec 06 16:49:00 crc kubenswrapper[4813]: I1206 16:49:00.249551 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jgn6l/crc-debug-4p6mj"] Dec 06 16:49:00 crc kubenswrapper[4813]: I1206 16:49:00.366589 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kh6v\" (UniqueName: \"kubernetes.io/projected/18497f1a-4db2-445a-9286-8a331fe18c30-kube-api-access-8kh6v\") pod \"18497f1a-4db2-445a-9286-8a331fe18c30\" (UID: \"18497f1a-4db2-445a-9286-8a331fe18c30\") " Dec 06 16:49:00 crc kubenswrapper[4813]: I1206 16:49:00.367390 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/18497f1a-4db2-445a-9286-8a331fe18c30-host\") pod \"18497f1a-4db2-445a-9286-8a331fe18c30\" (UID: \"18497f1a-4db2-445a-9286-8a331fe18c30\") " Dec 06 16:49:00 crc kubenswrapper[4813]: I1206 16:49:00.367517 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/18497f1a-4db2-445a-9286-8a331fe18c30-host" (OuterVolumeSpecName: "host") pod "18497f1a-4db2-445a-9286-8a331fe18c30" (UID: "18497f1a-4db2-445a-9286-8a331fe18c30"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 16:49:00 crc kubenswrapper[4813]: I1206 16:49:00.368512 4813 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/18497f1a-4db2-445a-9286-8a331fe18c30-host\") on node \"crc\" DevicePath \"\"" Dec 06 16:49:00 crc kubenswrapper[4813]: I1206 16:49:00.372275 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18497f1a-4db2-445a-9286-8a331fe18c30-kube-api-access-8kh6v" (OuterVolumeSpecName: "kube-api-access-8kh6v") pod "18497f1a-4db2-445a-9286-8a331fe18c30" (UID: "18497f1a-4db2-445a-9286-8a331fe18c30"). InnerVolumeSpecName "kube-api-access-8kh6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:49:00 crc kubenswrapper[4813]: I1206 16:49:00.470123 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kh6v\" (UniqueName: \"kubernetes.io/projected/18497f1a-4db2-445a-9286-8a331fe18c30-kube-api-access-8kh6v\") on node \"crc\" DevicePath \"\"" Dec 06 16:49:00 crc kubenswrapper[4813]: I1206 16:49:00.495700 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18497f1a-4db2-445a-9286-8a331fe18c30" path="/var/lib/kubelet/pods/18497f1a-4db2-445a-9286-8a331fe18c30/volumes" Dec 06 16:49:01 crc kubenswrapper[4813]: I1206 16:49:01.134531 4813 scope.go:117] "RemoveContainer" containerID="99f91363a0116b1813a6b42e4861f9e144ae96c33c70212ed502353488f63807" Dec 06 16:49:01 crc kubenswrapper[4813]: I1206 16:49:01.134536 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgn6l/crc-debug-4p6mj" Dec 06 16:49:01 crc kubenswrapper[4813]: I1206 16:49:01.422206 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jgn6l/crc-debug-sgknm"] Dec 06 16:49:01 crc kubenswrapper[4813]: E1206 16:49:01.422571 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18497f1a-4db2-445a-9286-8a331fe18c30" containerName="container-00" Dec 06 16:49:01 crc kubenswrapper[4813]: I1206 16:49:01.422583 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="18497f1a-4db2-445a-9286-8a331fe18c30" containerName="container-00" Dec 06 16:49:01 crc kubenswrapper[4813]: I1206 16:49:01.422753 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="18497f1a-4db2-445a-9286-8a331fe18c30" containerName="container-00" Dec 06 16:49:01 crc kubenswrapper[4813]: I1206 16:49:01.423315 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgn6l/crc-debug-sgknm" Dec 06 16:49:01 crc kubenswrapper[4813]: I1206 16:49:01.589185 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/15615b9e-4013-45e9-8dc0-506e0dabeb05-host\") pod \"crc-debug-sgknm\" (UID: \"15615b9e-4013-45e9-8dc0-506e0dabeb05\") " pod="openshift-must-gather-jgn6l/crc-debug-sgknm" Dec 06 16:49:01 crc kubenswrapper[4813]: I1206 16:49:01.589301 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlp4v\" (UniqueName: \"kubernetes.io/projected/15615b9e-4013-45e9-8dc0-506e0dabeb05-kube-api-access-hlp4v\") pod \"crc-debug-sgknm\" (UID: \"15615b9e-4013-45e9-8dc0-506e0dabeb05\") " pod="openshift-must-gather-jgn6l/crc-debug-sgknm" Dec 06 16:49:01 crc kubenswrapper[4813]: I1206 16:49:01.691399 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/15615b9e-4013-45e9-8dc0-506e0dabeb05-host\") pod \"crc-debug-sgknm\" (UID: \"15615b9e-4013-45e9-8dc0-506e0dabeb05\") " pod="openshift-must-gather-jgn6l/crc-debug-sgknm" Dec 06 16:49:01 crc kubenswrapper[4813]: I1206 16:49:01.691452 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlp4v\" (UniqueName: \"kubernetes.io/projected/15615b9e-4013-45e9-8dc0-506e0dabeb05-kube-api-access-hlp4v\") pod \"crc-debug-sgknm\" (UID: \"15615b9e-4013-45e9-8dc0-506e0dabeb05\") " pod="openshift-must-gather-jgn6l/crc-debug-sgknm" Dec 06 16:49:01 crc kubenswrapper[4813]: I1206 16:49:01.691871 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/15615b9e-4013-45e9-8dc0-506e0dabeb05-host\") pod \"crc-debug-sgknm\" (UID: \"15615b9e-4013-45e9-8dc0-506e0dabeb05\") " pod="openshift-must-gather-jgn6l/crc-debug-sgknm" Dec 06 16:49:01 crc kubenswrapper[4813]: I1206 16:49:01.708606 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlp4v\" (UniqueName: \"kubernetes.io/projected/15615b9e-4013-45e9-8dc0-506e0dabeb05-kube-api-access-hlp4v\") pod \"crc-debug-sgknm\" (UID: \"15615b9e-4013-45e9-8dc0-506e0dabeb05\") " pod="openshift-must-gather-jgn6l/crc-debug-sgknm" Dec 06 16:49:01 crc kubenswrapper[4813]: I1206 16:49:01.737727 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgn6l/crc-debug-sgknm" Dec 06 16:49:01 crc kubenswrapper[4813]: W1206 16:49:01.761578 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15615b9e_4013_45e9_8dc0_506e0dabeb05.slice/crio-c498cf527d1cae08a806228e3bd5b2a83dee397e5cd2cc4995480b3648e17060 WatchSource:0}: Error finding container c498cf527d1cae08a806228e3bd5b2a83dee397e5cd2cc4995480b3648e17060: Status 404 returned error can't find the container with id c498cf527d1cae08a806228e3bd5b2a83dee397e5cd2cc4995480b3648e17060 Dec 06 16:49:02 crc kubenswrapper[4813]: I1206 16:49:02.145195 4813 generic.go:334] "Generic (PLEG): container finished" podID="15615b9e-4013-45e9-8dc0-506e0dabeb05" containerID="ea6be8da46f9f5c1fd1da9081cec11e2851e2ff040f37645580d8f880f26a1a3" exitCode=0 Dec 06 16:49:02 crc kubenswrapper[4813]: I1206 16:49:02.145282 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jgn6l/crc-debug-sgknm" event={"ID":"15615b9e-4013-45e9-8dc0-506e0dabeb05","Type":"ContainerDied","Data":"ea6be8da46f9f5c1fd1da9081cec11e2851e2ff040f37645580d8f880f26a1a3"} Dec 06 16:49:02 crc kubenswrapper[4813]: I1206 16:49:02.145976 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jgn6l/crc-debug-sgknm" event={"ID":"15615b9e-4013-45e9-8dc0-506e0dabeb05","Type":"ContainerStarted","Data":"c498cf527d1cae08a806228e3bd5b2a83dee397e5cd2cc4995480b3648e17060"} Dec 06 16:49:02 crc kubenswrapper[4813]: I1206 16:49:02.177153 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jgn6l/crc-debug-sgknm"] Dec 06 16:49:02 crc kubenswrapper[4813]: I1206 16:49:02.184449 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jgn6l/crc-debug-sgknm"] Dec 06 16:49:03 crc kubenswrapper[4813]: I1206 16:49:03.282526 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgn6l/crc-debug-sgknm" Dec 06 16:49:03 crc kubenswrapper[4813]: I1206 16:49:03.341506 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/15615b9e-4013-45e9-8dc0-506e0dabeb05-host\") pod \"15615b9e-4013-45e9-8dc0-506e0dabeb05\" (UID: \"15615b9e-4013-45e9-8dc0-506e0dabeb05\") " Dec 06 16:49:03 crc kubenswrapper[4813]: I1206 16:49:03.341619 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlp4v\" (UniqueName: \"kubernetes.io/projected/15615b9e-4013-45e9-8dc0-506e0dabeb05-kube-api-access-hlp4v\") pod \"15615b9e-4013-45e9-8dc0-506e0dabeb05\" (UID: \"15615b9e-4013-45e9-8dc0-506e0dabeb05\") " Dec 06 16:49:03 crc kubenswrapper[4813]: I1206 16:49:03.341605 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15615b9e-4013-45e9-8dc0-506e0dabeb05-host" (OuterVolumeSpecName: "host") pod "15615b9e-4013-45e9-8dc0-506e0dabeb05" (UID: "15615b9e-4013-45e9-8dc0-506e0dabeb05"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 16:49:03 crc kubenswrapper[4813]: I1206 16:49:03.342320 4813 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/15615b9e-4013-45e9-8dc0-506e0dabeb05-host\") on node \"crc\" DevicePath \"\"" Dec 06 16:49:03 crc kubenswrapper[4813]: I1206 16:49:03.352446 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15615b9e-4013-45e9-8dc0-506e0dabeb05-kube-api-access-hlp4v" (OuterVolumeSpecName: "kube-api-access-hlp4v") pod "15615b9e-4013-45e9-8dc0-506e0dabeb05" (UID: "15615b9e-4013-45e9-8dc0-506e0dabeb05"). InnerVolumeSpecName "kube-api-access-hlp4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:49:03 crc kubenswrapper[4813]: I1206 16:49:03.444420 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlp4v\" (UniqueName: \"kubernetes.io/projected/15615b9e-4013-45e9-8dc0-506e0dabeb05-kube-api-access-hlp4v\") on node \"crc\" DevicePath \"\"" Dec 06 16:49:04 crc kubenswrapper[4813]: I1206 16:49:04.173966 4813 scope.go:117] "RemoveContainer" containerID="ea6be8da46f9f5c1fd1da9081cec11e2851e2ff040f37645580d8f880f26a1a3" Dec 06 16:49:04 crc kubenswrapper[4813]: I1206 16:49:04.174005 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgn6l/crc-debug-sgknm" Dec 06 16:49:04 crc kubenswrapper[4813]: I1206 16:49:04.507110 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15615b9e-4013-45e9-8dc0-506e0dabeb05" path="/var/lib/kubelet/pods/15615b9e-4013-45e9-8dc0-506e0dabeb05/volumes" Dec 06 16:49:06 crc kubenswrapper[4813]: I1206 16:49:06.486893 4813 scope.go:117] "RemoveContainer" containerID="969c7429b9faa2222909b4bad4ef827351b31f2cde6d6ab68e5af156f1ab0dfc" Dec 06 16:49:06 crc kubenswrapper[4813]: E1206 16:49:06.487691 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:49:21 crc kubenswrapper[4813]: I1206 16:49:21.487447 4813 scope.go:117] "RemoveContainer" containerID="969c7429b9faa2222909b4bad4ef827351b31f2cde6d6ab68e5af156f1ab0dfc" Dec 06 16:49:21 crc kubenswrapper[4813]: E1206 16:49:21.488236 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:49:31 crc kubenswrapper[4813]: I1206 16:49:31.035686 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-58d4d548d-6bl8w_f883aaa3-f010-4c52-bc12-1591fb0ddb4d/barbican-api/0.log" Dec 06 16:49:31 crc kubenswrapper[4813]: I1206 16:49:31.368465 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-64449f75dd-k89r7_322df0dd-86e6-466b-8afa-e4e52e889827/barbican-keystone-listener/0.log" Dec 06 16:49:31 crc kubenswrapper[4813]: I1206 16:49:31.589207 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-ffc86b7c-26pxs_cf88c989-e78b-456e-ba32-19d8af98a1d0/barbican-worker/0.log" Dec 06 16:49:31 crc kubenswrapper[4813]: I1206 16:49:31.600592 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-64449f75dd-k89r7_322df0dd-86e6-466b-8afa-e4e52e889827/barbican-keystone-listener-log/0.log" Dec 06 16:49:31 crc kubenswrapper[4813]: I1206 16:49:31.756802 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-58d4d548d-6bl8w_f883aaa3-f010-4c52-bc12-1591fb0ddb4d/barbican-api-log/0.log" Dec 06 16:49:31 crc kubenswrapper[4813]: I1206 16:49:31.791913 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-ffc86b7c-26pxs_cf88c989-e78b-456e-ba32-19d8af98a1d0/barbican-worker-log/0.log" Dec 06 16:49:32 crc kubenswrapper[4813]: I1206 16:49:32.108053 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-qr892_6306b1cf-bf21-4cd0-a7c8-c83a99863266/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:49:32 crc kubenswrapper[4813]: I1206 16:49:32.192709 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_66613ee6-b0b7-48fc-adb6-86bab4940550/ceilometer-central-agent/0.log" Dec 06 16:49:32 crc kubenswrapper[4813]: I1206 16:49:32.320724 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_66613ee6-b0b7-48fc-adb6-86bab4940550/proxy-httpd/0.log" Dec 06 16:49:32 crc kubenswrapper[4813]: I1206 16:49:32.363225 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_66613ee6-b0b7-48fc-adb6-86bab4940550/sg-core/0.log" Dec 06 16:49:32 crc kubenswrapper[4813]: I1206 16:49:32.374379 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_66613ee6-b0b7-48fc-adb6-86bab4940550/ceilometer-notification-agent/0.log" Dec 06 16:49:32 crc kubenswrapper[4813]: I1206 16:49:32.490404 4813 scope.go:117] "RemoveContainer" containerID="969c7429b9faa2222909b4bad4ef827351b31f2cde6d6ab68e5af156f1ab0dfc" Dec 06 16:49:32 crc kubenswrapper[4813]: E1206 16:49:32.490706 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:49:32 crc kubenswrapper[4813]: I1206 16:49:32.558633 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d/cinder-api-log/0.log" Dec 06 16:49:32 crc kubenswrapper[4813]: I1206 16:49:32.594199 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_c36fcd92-f5a1-4ec9-a1ee-fa6b54522e5d/cinder-api/0.log" Dec 06 16:49:32 crc kubenswrapper[4813]: I1206 16:49:32.789655 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_7d707d75-8ac3-4256-9e9e-018e5caef917/cinder-scheduler/0.log" Dec 06 16:49:32 crc kubenswrapper[4813]: I1206 16:49:32.870376 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_7d707d75-8ac3-4256-9e9e-018e5caef917/probe/0.log" Dec 06 16:49:32 crc kubenswrapper[4813]: I1206 16:49:32.902510 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-48kg5_11276539-afd1-4649-97eb-867dfcc76819/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:49:33 crc kubenswrapper[4813]: I1206 16:49:33.106384 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-gfhcl_6f525ab6-f23e-4cc9-a843-7a601896c864/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:49:33 crc kubenswrapper[4813]: I1206 16:49:33.234431 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6b6dc74c5-4w85d_5399f130-d1b7-48ba-a32b-6da83416719e/init/0.log" Dec 06 16:49:33 crc kubenswrapper[4813]: I1206 16:49:33.393390 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6b6dc74c5-4w85d_5399f130-d1b7-48ba-a32b-6da83416719e/init/0.log" Dec 06 16:49:33 crc kubenswrapper[4813]: I1206 16:49:33.444511 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6b6dc74c5-4w85d_5399f130-d1b7-48ba-a32b-6da83416719e/dnsmasq-dns/0.log" Dec 06 16:49:33 crc kubenswrapper[4813]: I1206 16:49:33.545857 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-pkk2h_c2bb7ef8-9c22-412d-83a9-686036ecb3f8/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:49:33 crc kubenswrapper[4813]: I1206 16:49:33.685998 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_ee3e963f-6982-4cd8-87d2-4eceaec10be3/glance-log/0.log" Dec 06 16:49:33 crc kubenswrapper[4813]: I1206 16:49:33.750438 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_ee3e963f-6982-4cd8-87d2-4eceaec10be3/glance-httpd/0.log" Dec 06 16:49:33 crc kubenswrapper[4813]: I1206 16:49:33.900313 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_ec17345c-3d8e-4981-945c-173fb39aab99/glance-httpd/0.log" Dec 06 16:49:33 crc kubenswrapper[4813]: I1206 16:49:33.970303 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_ec17345c-3d8e-4981-945c-173fb39aab99/glance-log/0.log" Dec 06 16:49:34 crc kubenswrapper[4813]: I1206 16:49:34.114394 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7fd6f7946b-kmm96_28238998-5d44-4ae9-8d1f-93d56ff18152/horizon/0.log" Dec 06 16:49:34 crc kubenswrapper[4813]: I1206 16:49:34.367232 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-jjhcg_8c3bde1a-68be-4aaf-9c46-f5ea14578783/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:49:34 crc kubenswrapper[4813]: I1206 16:49:34.441926 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7fd6f7946b-kmm96_28238998-5d44-4ae9-8d1f-93d56ff18152/horizon-log/0.log" Dec 06 16:49:34 crc kubenswrapper[4813]: I1206 16:49:34.567903 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-nbw6k_c8332abd-e13b-4587-831b-1330e18de573/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:49:34 crc kubenswrapper[4813]: I1206 16:49:34.811912 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-c4f8d8d9b-p5pqq_c6d636bb-c76a-4895-8af6-6801f45e4e5b/keystone-api/0.log" Dec 06 16:49:34 crc kubenswrapper[4813]: I1206 16:49:34.858637 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_8d3a2b2d-c016-4948-83c5-f8fcae8678d8/kube-state-metrics/0.log" Dec 06 16:49:35 crc kubenswrapper[4813]: I1206 16:49:35.095003 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-h5v4j_bfa11078-d153-43a2-abc4-b70f491dc4a6/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:49:35 crc kubenswrapper[4813]: I1206 16:49:35.338946 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-759b8b45fc-psf25_afd03411-dfea-469b-804c-7cbe8a964d54/neutron-api/0.log" Dec 06 16:49:35 crc kubenswrapper[4813]: I1206 16:49:35.473474 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-759b8b45fc-psf25_afd03411-dfea-469b-804c-7cbe8a964d54/neutron-httpd/0.log" Dec 06 16:49:35 crc kubenswrapper[4813]: I1206 16:49:35.570636 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-7vrmk_a11c7f97-21d3-48b1-9640-529df9d946b8/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:49:35 crc kubenswrapper[4813]: I1206 16:49:35.929944 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_e54ec1f1-806c-475d-843b-53a55fea0fab/nova-api-log/0.log" Dec 06 16:49:36 crc kubenswrapper[4813]: I1206 16:49:36.078716 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_db38328d-a236-4b36-bb91-9a9b818d39f1/nova-cell0-conductor-conductor/0.log" Dec 06 16:49:36 crc kubenswrapper[4813]: I1206 16:49:36.317029 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_e54ec1f1-806c-475d-843b-53a55fea0fab/nova-api-api/0.log" Dec 06 16:49:36 crc kubenswrapper[4813]: I1206 16:49:36.396438 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_0a45c584-458c-4f14-b7f5-50711d94207c/nova-cell1-conductor-conductor/0.log" Dec 06 16:49:37 crc kubenswrapper[4813]: I1206 16:49:37.105371 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_e9d64e47-211f-4eab-84e9-0caabeff895b/nova-cell1-novncproxy-novncproxy/0.log" Dec 06 16:49:37 crc kubenswrapper[4813]: I1206 16:49:37.204756 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-mfh5s_7ba82e65-8192-4943-8a2a-863d041ad231/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:49:37 crc kubenswrapper[4813]: I1206 16:49:37.419369 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_1731c7e7-8ded-4a61-8b8a-d94cf8b4a860/nova-metadata-log/0.log" Dec 06 16:49:37 crc kubenswrapper[4813]: I1206 16:49:37.725793 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_d53150b5-8980-47ff-bc3a-6b75ac75fa14/nova-scheduler-scheduler/0.log" Dec 06 16:49:37 crc kubenswrapper[4813]: I1206 16:49:37.803378 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_bf7377f8-ef23-484a-b66b-90db1cd5b7cd/mysql-bootstrap/0.log" Dec 06 16:49:38 crc kubenswrapper[4813]: I1206 16:49:38.093978 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_bf7377f8-ef23-484a-b66b-90db1cd5b7cd/mysql-bootstrap/0.log" Dec 06 16:49:38 crc kubenswrapper[4813]: I1206 16:49:38.115487 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_bf7377f8-ef23-484a-b66b-90db1cd5b7cd/galera/0.log" Dec 06 16:49:38 crc kubenswrapper[4813]: I1206 16:49:38.632005 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_75d94fff-d912-4493-b9b9-c458138a7ccf/mysql-bootstrap/0.log" Dec 06 16:49:38 crc kubenswrapper[4813]: I1206 16:49:38.728834 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_1731c7e7-8ded-4a61-8b8a-d94cf8b4a860/nova-metadata-metadata/0.log" Dec 06 16:49:38 crc kubenswrapper[4813]: I1206 16:49:38.895222 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_75d94fff-d912-4493-b9b9-c458138a7ccf/mysql-bootstrap/0.log" Dec 06 16:49:38 crc kubenswrapper[4813]: I1206 16:49:38.938973 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_75d94fff-d912-4493-b9b9-c458138a7ccf/galera/0.log" Dec 06 16:49:38 crc kubenswrapper[4813]: I1206 16:49:38.976410 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_0447affb-1196-4c65-8e3d-936864d4fdba/openstackclient/0.log" Dec 06 16:49:39 crc kubenswrapper[4813]: I1206 16:49:39.202296 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-bg6cb_ba6825c1-2137-46d6-a695-6bdd53a13ec4/ovn-controller/0.log" Dec 06 16:49:39 crc kubenswrapper[4813]: I1206 16:49:39.224188 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-9bcrr_da431468-e795-477e-97cd-b1796a76a117/openstack-network-exporter/0.log" Dec 06 16:49:39 crc kubenswrapper[4813]: I1206 16:49:39.429180 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mk4nr_6b7cce06-d3db-47cd-a542-a77bfdc42f82/ovsdb-server-init/0.log" Dec 06 16:49:39 crc kubenswrapper[4813]: I1206 16:49:39.676687 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mk4nr_6b7cce06-d3db-47cd-a542-a77bfdc42f82/ovs-vswitchd/0.log" Dec 06 16:49:39 crc kubenswrapper[4813]: I1206 16:49:39.684879 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mk4nr_6b7cce06-d3db-47cd-a542-a77bfdc42f82/ovsdb-server-init/0.log" Dec 06 16:49:39 crc kubenswrapper[4813]: I1206 16:49:39.749029 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mk4nr_6b7cce06-d3db-47cd-a542-a77bfdc42f82/ovsdb-server/0.log" Dec 06 16:49:39 crc kubenswrapper[4813]: I1206 16:49:39.967747 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-6gtnd_ef88d823-7e10-4eee-a6b1-66c680355906/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:49:40 crc kubenswrapper[4813]: I1206 16:49:40.015634 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_9614929b-2a14-4d49-8009-8f627fb000fd/openstack-network-exporter/0.log" Dec 06 16:49:40 crc kubenswrapper[4813]: I1206 16:49:40.223750 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_9614929b-2a14-4d49-8009-8f627fb000fd/ovn-northd/0.log" Dec 06 16:49:40 crc kubenswrapper[4813]: I1206 16:49:40.238224 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c0a272bd-6444-4c90-acb0-fb653bf3b180/openstack-network-exporter/0.log" Dec 06 16:49:40 crc kubenswrapper[4813]: I1206 16:49:40.238376 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c0a272bd-6444-4c90-acb0-fb653bf3b180/ovsdbserver-nb/0.log" Dec 06 16:49:40 crc kubenswrapper[4813]: I1206 16:49:40.474679 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420/ovsdbserver-sb/0.log" Dec 06 16:49:40 crc kubenswrapper[4813]: I1206 16:49:40.521231 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7cce2bf2-3398-47a8-8d0d-ae8ef8ebd420/openstack-network-exporter/0.log" Dec 06 16:49:40 crc kubenswrapper[4813]: I1206 16:49:40.797483 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-98dddfffb-l698f_e913dfcd-dfe3-4c2c-b72a-8393295c95fc/placement-api/0.log" Dec 06 16:49:40 crc kubenswrapper[4813]: I1206 16:49:40.818956 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-98dddfffb-l698f_e913dfcd-dfe3-4c2c-b72a-8393295c95fc/placement-log/0.log" Dec 06 16:49:40 crc kubenswrapper[4813]: I1206 16:49:40.861786 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_342df504-7a38-4c50-bb3d-8c7b28521e64/setup-container/0.log" Dec 06 16:49:41 crc kubenswrapper[4813]: I1206 16:49:41.114637 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_342df504-7a38-4c50-bb3d-8c7b28521e64/setup-container/0.log" Dec 06 16:49:41 crc kubenswrapper[4813]: I1206 16:49:41.129506 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_342df504-7a38-4c50-bb3d-8c7b28521e64/rabbitmq/0.log" Dec 06 16:49:41 crc kubenswrapper[4813]: I1206 16:49:41.151467 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_5ef242c1-0127-4963-9563-4943a584398c/setup-container/0.log" Dec 06 16:49:41 crc kubenswrapper[4813]: I1206 16:49:41.378057 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_5ef242c1-0127-4963-9563-4943a584398c/rabbitmq/0.log" Dec 06 16:49:41 crc kubenswrapper[4813]: I1206 16:49:41.384060 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-qpwx5_0ddaea40-3167-4f8b-8de0-756048946c8c/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:49:41 crc kubenswrapper[4813]: I1206 16:49:41.421722 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_5ef242c1-0127-4963-9563-4943a584398c/setup-container/0.log" Dec 06 16:49:41 crc kubenswrapper[4813]: I1206 16:49:41.650041 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-n6p6l_aa40ab8b-bc66-4e51-a126-ec15a6250486/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:49:41 crc kubenswrapper[4813]: I1206 16:49:41.664720 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-p9rg2_ef5432f4-56a6-4518-a6b1-00ac9b29293f/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:49:41 crc kubenswrapper[4813]: I1206 16:49:41.929757 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-h7rjq_97035b28-6546-4c6b-b273-085f7ca5bb4f/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:49:41 crc kubenswrapper[4813]: I1206 16:49:41.981056 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-nj5n7_41e561aa-3484-4946-a8d9-174ea9a22a45/ssh-known-hosts-edpm-deployment/0.log" Dec 06 16:49:42 crc kubenswrapper[4813]: I1206 16:49:42.237011 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-85cc9c6459-lnvlp_c4b18e9f-80a6-480e-878b-3acf0a5716a3/proxy-server/0.log" Dec 06 16:49:42 crc kubenswrapper[4813]: I1206 16:49:42.379063 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-7pplm_f67c4c64-2c67-46bc-af93-6ca6ac4f4814/swift-ring-rebalance/0.log" Dec 06 16:49:42 crc kubenswrapper[4813]: I1206 16:49:42.432608 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-85cc9c6459-lnvlp_c4b18e9f-80a6-480e-878b-3acf0a5716a3/proxy-httpd/0.log" Dec 06 16:49:42 crc kubenswrapper[4813]: I1206 16:49:42.550079 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f571761-c8de-485f-89de-514fc0a66a55/account-auditor/0.log" Dec 06 16:49:42 crc kubenswrapper[4813]: I1206 16:49:42.625001 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f571761-c8de-485f-89de-514fc0a66a55/account-reaper/0.log" Dec 06 16:49:42 crc kubenswrapper[4813]: I1206 16:49:42.699460 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f571761-c8de-485f-89de-514fc0a66a55/account-replicator/0.log" Dec 06 16:49:42 crc kubenswrapper[4813]: I1206 16:49:42.751118 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f571761-c8de-485f-89de-514fc0a66a55/account-server/0.log" Dec 06 16:49:42 crc kubenswrapper[4813]: I1206 16:49:42.820425 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f571761-c8de-485f-89de-514fc0a66a55/container-auditor/0.log" Dec 06 16:49:42 crc kubenswrapper[4813]: I1206 16:49:42.904018 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f571761-c8de-485f-89de-514fc0a66a55/container-server/0.log" Dec 06 16:49:42 crc kubenswrapper[4813]: I1206 16:49:42.916623 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f571761-c8de-485f-89de-514fc0a66a55/container-replicator/0.log" Dec 06 16:49:42 crc kubenswrapper[4813]: I1206 16:49:42.999168 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f571761-c8de-485f-89de-514fc0a66a55/container-updater/0.log" Dec 06 16:49:43 crc kubenswrapper[4813]: I1206 16:49:43.101905 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f571761-c8de-485f-89de-514fc0a66a55/object-expirer/0.log" Dec 06 16:49:43 crc kubenswrapper[4813]: I1206 16:49:43.139119 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f571761-c8de-485f-89de-514fc0a66a55/object-auditor/0.log" Dec 06 16:49:43 crc kubenswrapper[4813]: I1206 16:49:43.265479 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f571761-c8de-485f-89de-514fc0a66a55/object-replicator/0.log" Dec 06 16:49:43 crc kubenswrapper[4813]: I1206 16:49:43.288550 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f571761-c8de-485f-89de-514fc0a66a55/object-server/0.log" Dec 06 16:49:43 crc kubenswrapper[4813]: I1206 16:49:43.388290 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f571761-c8de-485f-89de-514fc0a66a55/object-updater/0.log" Dec 06 16:49:43 crc kubenswrapper[4813]: I1206 16:49:43.441433 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f571761-c8de-485f-89de-514fc0a66a55/rsync/0.log" Dec 06 16:49:43 crc kubenswrapper[4813]: I1206 16:49:43.517832 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f571761-c8de-485f-89de-514fc0a66a55/swift-recon-cron/0.log" Dec 06 16:49:43 crc kubenswrapper[4813]: I1206 16:49:43.730359 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-zqrmt_a1ecf1f7-01c3-456c-8698-1353e0651c24/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:49:43 crc kubenswrapper[4813]: I1206 16:49:43.843585 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_5e4bd588-0ab4-4132-a2e8-1c5ea769f352/tempest-tests-tempest-tests-runner/0.log" Dec 06 16:49:43 crc kubenswrapper[4813]: I1206 16:49:43.935108 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_9899c9cb-fe20-41a1-a200-f70d5b63b314/test-operator-logs-container/0.log" Dec 06 16:49:44 crc kubenswrapper[4813]: I1206 16:49:44.338988 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-7tjp5_07e3bef8-6975-428b-ac2e-265d2db3d70f/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 16:49:44 crc kubenswrapper[4813]: I1206 16:49:44.488659 4813 scope.go:117] "RemoveContainer" containerID="969c7429b9faa2222909b4bad4ef827351b31f2cde6d6ab68e5af156f1ab0dfc" Dec 06 16:49:44 crc kubenswrapper[4813]: E1206 16:49:44.488888 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:49:55 crc kubenswrapper[4813]: I1206 16:49:55.660048 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_9fab2eee-6b9d-4fd2-b3cc-0468fed71b2c/memcached/0.log" Dec 06 16:49:59 crc kubenswrapper[4813]: I1206 16:49:59.487143 4813 scope.go:117] "RemoveContainer" containerID="969c7429b9faa2222909b4bad4ef827351b31f2cde6d6ab68e5af156f1ab0dfc" Dec 06 16:49:59 crc kubenswrapper[4813]: E1206 16:49:59.487653 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:50:14 crc kubenswrapper[4813]: I1206 16:50:14.487298 4813 scope.go:117] "RemoveContainer" containerID="969c7429b9faa2222909b4bad4ef827351b31f2cde6d6ab68e5af156f1ab0dfc" Dec 06 16:50:14 crc kubenswrapper[4813]: E1206 16:50:14.488069 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:50:15 crc kubenswrapper[4813]: I1206 16:50:15.015053 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f_89c13d20-79f6-4a15-9c75-cf199679fc7e/util/0.log" Dec 06 16:50:15 crc kubenswrapper[4813]: I1206 16:50:15.256503 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f_89c13d20-79f6-4a15-9c75-cf199679fc7e/util/0.log" Dec 06 16:50:15 crc kubenswrapper[4813]: I1206 16:50:15.279845 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f_89c13d20-79f6-4a15-9c75-cf199679fc7e/pull/0.log" Dec 06 16:50:15 crc kubenswrapper[4813]: I1206 16:50:15.279919 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f_89c13d20-79f6-4a15-9c75-cf199679fc7e/pull/0.log" Dec 06 16:50:15 crc kubenswrapper[4813]: I1206 16:50:15.551122 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f_89c13d20-79f6-4a15-9c75-cf199679fc7e/util/0.log" Dec 06 16:50:15 crc kubenswrapper[4813]: I1206 16:50:15.555487 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f_89c13d20-79f6-4a15-9c75-cf199679fc7e/extract/0.log" Dec 06 16:50:15 crc kubenswrapper[4813]: I1206 16:50:15.612937 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_05f94be2844c7788072bae73296f732928c34a06a9b0982fae500affbcnrh7f_89c13d20-79f6-4a15-9c75-cf199679fc7e/pull/0.log" Dec 06 16:50:15 crc kubenswrapper[4813]: I1206 16:50:15.775570 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6dc7dc95b4-6wznz_24d6dd0f-86e6-4b7a-beeb-56025a76ab5a/kube-rbac-proxy/0.log" Dec 06 16:50:15 crc kubenswrapper[4813]: I1206 16:50:15.834914 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6dc7dc95b4-6wznz_24d6dd0f-86e6-4b7a-beeb-56025a76ab5a/manager/0.log" Dec 06 16:50:15 crc kubenswrapper[4813]: I1206 16:50:15.854525 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-8mt55_a8b0c607-9997-44e8-a3cc-a36e4b2e70d2/kube-rbac-proxy/0.log" Dec 06 16:50:16 crc kubenswrapper[4813]: I1206 16:50:16.053751 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-8mt55_a8b0c607-9997-44e8-a3cc-a36e4b2e70d2/manager/0.log" Dec 06 16:50:16 crc kubenswrapper[4813]: I1206 16:50:16.084108 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-4rdzx_218be173-f41f-48b1-8af0-b94bbc483e8b/kube-rbac-proxy/0.log" Dec 06 16:50:16 crc kubenswrapper[4813]: I1206 16:50:16.111863 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-4rdzx_218be173-f41f-48b1-8af0-b94bbc483e8b/manager/0.log" Dec 06 16:50:16 crc kubenswrapper[4813]: I1206 16:50:16.359764 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-f9nfd_3edd7ca8-a3ba-46c5-96c8-77aa9be672b7/kube-rbac-proxy/0.log" Dec 06 16:50:16 crc kubenswrapper[4813]: I1206 16:50:16.424394 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-f9nfd_3edd7ca8-a3ba-46c5-96c8-77aa9be672b7/manager/0.log" Dec 06 16:50:16 crc kubenswrapper[4813]: I1206 16:50:16.501881 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-wxsqx_e33ae276-0e3d-44b5-b970-17e98051c49c/kube-rbac-proxy/0.log" Dec 06 16:50:16 crc kubenswrapper[4813]: I1206 16:50:16.564181 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-wxsqx_e33ae276-0e3d-44b5-b970-17e98051c49c/manager/0.log" Dec 06 16:50:16 crc kubenswrapper[4813]: I1206 16:50:16.666653 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-g4ztj_d67349e5-13fb-4859-ac8d-3e2c6bd67dbb/kube-rbac-proxy/0.log" Dec 06 16:50:16 crc kubenswrapper[4813]: I1206 16:50:16.759499 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-g4ztj_d67349e5-13fb-4859-ac8d-3e2c6bd67dbb/manager/0.log" Dec 06 16:50:16 crc kubenswrapper[4813]: I1206 16:50:16.855708 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-pkvxj_6983648f-60ef-429d-a2fd-de5dc7e6d0ba/kube-rbac-proxy/0.log" Dec 06 16:50:17 crc kubenswrapper[4813]: I1206 16:50:17.096434 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-cxh72_4d2f63b5-c93e-4949-bb79-459dad74dcf8/manager/0.log" Dec 06 16:50:17 crc kubenswrapper[4813]: I1206 16:50:17.104355 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-pkvxj_6983648f-60ef-429d-a2fd-de5dc7e6d0ba/manager/0.log" Dec 06 16:50:17 crc kubenswrapper[4813]: I1206 16:50:17.122895 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-cxh72_4d2f63b5-c93e-4949-bb79-459dad74dcf8/kube-rbac-proxy/0.log" Dec 06 16:50:17 crc kubenswrapper[4813]: I1206 16:50:17.307494 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-cjzht_8870fd51-2141-4cc8-956b-050e849dcbd4/kube-rbac-proxy/0.log" Dec 06 16:50:17 crc kubenswrapper[4813]: I1206 16:50:17.370196 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-cjzht_8870fd51-2141-4cc8-956b-050e849dcbd4/manager/0.log" Dec 06 16:50:17 crc kubenswrapper[4813]: I1206 16:50:17.455707 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-6p6sf_a8763d41-9404-43ec-866b-9244bf2c4af9/manager/0.log" Dec 06 16:50:17 crc kubenswrapper[4813]: I1206 16:50:17.509331 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-6p6sf_a8763d41-9404-43ec-866b-9244bf2c4af9/kube-rbac-proxy/0.log" Dec 06 16:50:17 crc kubenswrapper[4813]: I1206 16:50:17.584571 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-6n2hj_99a0eae7-c990-42cd-a18e-78b81177bbb6/kube-rbac-proxy/0.log" Dec 06 16:50:17 crc kubenswrapper[4813]: I1206 16:50:17.689557 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-6n2hj_99a0eae7-c990-42cd-a18e-78b81177bbb6/manager/0.log" Dec 06 16:50:17 crc kubenswrapper[4813]: I1206 16:50:17.808958 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-6rr2s_b4cd1114-c317-4351-b566-05317dd589f1/kube-rbac-proxy/0.log" Dec 06 16:50:17 crc kubenswrapper[4813]: I1206 16:50:17.925736 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-6rr2s_b4cd1114-c317-4351-b566-05317dd589f1/manager/0.log" Dec 06 16:50:18 crc kubenswrapper[4813]: I1206 16:50:18.030813 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-78fv6_b675acf0-51cc-43d9-b9c6-bd0ee1730d25/kube-rbac-proxy/0.log" Dec 06 16:50:18 crc kubenswrapper[4813]: I1206 16:50:18.057081 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-78fv6_b675acf0-51cc-43d9-b9c6-bd0ee1730d25/manager/0.log" Dec 06 16:50:18 crc kubenswrapper[4813]: I1206 16:50:18.206883 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-g9q96_031a9bd1-298a-4945-a52d-452005c8467e/kube-rbac-proxy/0.log" Dec 06 16:50:18 crc kubenswrapper[4813]: I1206 16:50:18.267291 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-g9q96_031a9bd1-298a-4945-a52d-452005c8467e/manager/0.log" Dec 06 16:50:18 crc kubenswrapper[4813]: I1206 16:50:18.362813 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879fc7jrk_279a2707-03fd-4cbc-9e13-d7a0d2a284db/kube-rbac-proxy/0.log" Dec 06 16:50:18 crc kubenswrapper[4813]: I1206 16:50:18.434038 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879fc7jrk_279a2707-03fd-4cbc-9e13-d7a0d2a284db/manager/0.log" Dec 06 16:50:18 crc kubenswrapper[4813]: I1206 16:50:18.901365 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-cb74cdcd8-r5tfz_d2bd6dc9-ba04-4168-b879-415d9ce263d4/operator/0.log" Dec 06 16:50:18 crc kubenswrapper[4813]: I1206 16:50:18.934012 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-vgjq7_74a02e6c-0417-4c92-9edd-18925cfbf96e/registry-server/0.log" Dec 06 16:50:19 crc kubenswrapper[4813]: I1206 16:50:19.270041 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-2mpwn_c35efae6-9bf9-466d-ac8d-203d8a2c4b97/manager/0.log" Dec 06 16:50:19 crc kubenswrapper[4813]: I1206 16:50:19.287496 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-2mpwn_c35efae6-9bf9-466d-ac8d-203d8a2c4b97/kube-rbac-proxy/0.log" Dec 06 16:50:19 crc kubenswrapper[4813]: I1206 16:50:19.410412 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5c8695d769-78g26_572ce1d1-6138-4f8e-8da9-c22a3b6b3c2b/manager/0.log" Dec 06 16:50:19 crc kubenswrapper[4813]: I1206 16:50:19.415010 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-rsbqf_67c6b7ea-97cf-4ffa-9f3b-5111d4fbdeba/kube-rbac-proxy/0.log" Dec 06 16:50:19 crc kubenswrapper[4813]: I1206 16:50:19.679085 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-rsbqf_67c6b7ea-97cf-4ffa-9f3b-5111d4fbdeba/manager/0.log" Dec 06 16:50:19 crc kubenswrapper[4813]: I1206 16:50:19.821463 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-9lwhh_00cfc20a-68d6-46be-98b8-83a43c9d11d9/operator/0.log" Dec 06 16:50:19 crc kubenswrapper[4813]: I1206 16:50:19.915352 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-x8vkl_b05dafaa-0d81-4b80-ad3e-48da48a1655f/kube-rbac-proxy/0.log" Dec 06 16:50:19 crc kubenswrapper[4813]: I1206 16:50:19.977902 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-x8vkl_b05dafaa-0d81-4b80-ad3e-48da48a1655f/manager/0.log" Dec 06 16:50:20 crc kubenswrapper[4813]: I1206 16:50:20.047734 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-qkrqw_6d54c577-3fdf-48af-9070-301dd3b3eb81/kube-rbac-proxy/0.log" Dec 06 16:50:20 crc kubenswrapper[4813]: I1206 16:50:20.166345 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-cbmwf_28acd95a-e657-46b5-a771-a24b27bdc6fe/kube-rbac-proxy/0.log" Dec 06 16:50:20 crc kubenswrapper[4813]: I1206 16:50:20.198540 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-qkrqw_6d54c577-3fdf-48af-9070-301dd3b3eb81/manager/0.log" Dec 06 16:50:20 crc kubenswrapper[4813]: I1206 16:50:20.241615 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-cbmwf_28acd95a-e657-46b5-a771-a24b27bdc6fe/manager/0.log" Dec 06 16:50:20 crc kubenswrapper[4813]: I1206 16:50:20.366411 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-8zvf5_5a8652b7-8a55-4658-acce-b7573421b94d/kube-rbac-proxy/0.log" Dec 06 16:50:20 crc kubenswrapper[4813]: I1206 16:50:20.424739 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-8zvf5_5a8652b7-8a55-4658-acce-b7573421b94d/manager/0.log" Dec 06 16:50:28 crc kubenswrapper[4813]: I1206 16:50:28.487073 4813 scope.go:117] "RemoveContainer" containerID="969c7429b9faa2222909b4bad4ef827351b31f2cde6d6ab68e5af156f1ab0dfc" Dec 06 16:50:28 crc kubenswrapper[4813]: E1206 16:50:28.488997 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:50:35 crc kubenswrapper[4813]: I1206 16:50:35.943200 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-b4wzv"] Dec 06 16:50:35 crc kubenswrapper[4813]: E1206 16:50:35.947972 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15615b9e-4013-45e9-8dc0-506e0dabeb05" containerName="container-00" Dec 06 16:50:35 crc kubenswrapper[4813]: I1206 16:50:35.948009 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="15615b9e-4013-45e9-8dc0-506e0dabeb05" containerName="container-00" Dec 06 16:50:35 crc kubenswrapper[4813]: I1206 16:50:35.948253 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="15615b9e-4013-45e9-8dc0-506e0dabeb05" containerName="container-00" Dec 06 16:50:35 crc kubenswrapper[4813]: I1206 16:50:35.950090 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b4wzv" Dec 06 16:50:35 crc kubenswrapper[4813]: I1206 16:50:35.974467 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b4wzv"] Dec 06 16:50:36 crc kubenswrapper[4813]: I1206 16:50:36.041191 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e6f68b4-913b-4d69-9c46-3e3e5f7634e1-utilities\") pod \"certified-operators-b4wzv\" (UID: \"9e6f68b4-913b-4d69-9c46-3e3e5f7634e1\") " pod="openshift-marketplace/certified-operators-b4wzv" Dec 06 16:50:36 crc kubenswrapper[4813]: I1206 16:50:36.041377 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtjsx\" (UniqueName: \"kubernetes.io/projected/9e6f68b4-913b-4d69-9c46-3e3e5f7634e1-kube-api-access-mtjsx\") pod \"certified-operators-b4wzv\" (UID: \"9e6f68b4-913b-4d69-9c46-3e3e5f7634e1\") " pod="openshift-marketplace/certified-operators-b4wzv" Dec 06 16:50:36 crc kubenswrapper[4813]: I1206 16:50:36.041417 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e6f68b4-913b-4d69-9c46-3e3e5f7634e1-catalog-content\") pod \"certified-operators-b4wzv\" (UID: \"9e6f68b4-913b-4d69-9c46-3e3e5f7634e1\") " pod="openshift-marketplace/certified-operators-b4wzv" Dec 06 16:50:36 crc kubenswrapper[4813]: I1206 16:50:36.143119 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtjsx\" (UniqueName: \"kubernetes.io/projected/9e6f68b4-913b-4d69-9c46-3e3e5f7634e1-kube-api-access-mtjsx\") pod \"certified-operators-b4wzv\" (UID: \"9e6f68b4-913b-4d69-9c46-3e3e5f7634e1\") " pod="openshift-marketplace/certified-operators-b4wzv" Dec 06 16:50:36 crc kubenswrapper[4813]: I1206 16:50:36.143205 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e6f68b4-913b-4d69-9c46-3e3e5f7634e1-catalog-content\") pod \"certified-operators-b4wzv\" (UID: \"9e6f68b4-913b-4d69-9c46-3e3e5f7634e1\") " pod="openshift-marketplace/certified-operators-b4wzv" Dec 06 16:50:36 crc kubenswrapper[4813]: I1206 16:50:36.143282 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e6f68b4-913b-4d69-9c46-3e3e5f7634e1-utilities\") pod \"certified-operators-b4wzv\" (UID: \"9e6f68b4-913b-4d69-9c46-3e3e5f7634e1\") " pod="openshift-marketplace/certified-operators-b4wzv" Dec 06 16:50:36 crc kubenswrapper[4813]: I1206 16:50:36.144095 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e6f68b4-913b-4d69-9c46-3e3e5f7634e1-utilities\") pod \"certified-operators-b4wzv\" (UID: \"9e6f68b4-913b-4d69-9c46-3e3e5f7634e1\") " pod="openshift-marketplace/certified-operators-b4wzv" Dec 06 16:50:36 crc kubenswrapper[4813]: I1206 16:50:36.144133 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e6f68b4-913b-4d69-9c46-3e3e5f7634e1-catalog-content\") pod \"certified-operators-b4wzv\" (UID: \"9e6f68b4-913b-4d69-9c46-3e3e5f7634e1\") " pod="openshift-marketplace/certified-operators-b4wzv" Dec 06 16:50:36 crc kubenswrapper[4813]: I1206 16:50:36.162528 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtjsx\" (UniqueName: \"kubernetes.io/projected/9e6f68b4-913b-4d69-9c46-3e3e5f7634e1-kube-api-access-mtjsx\") pod \"certified-operators-b4wzv\" (UID: \"9e6f68b4-913b-4d69-9c46-3e3e5f7634e1\") " pod="openshift-marketplace/certified-operators-b4wzv" Dec 06 16:50:36 crc kubenswrapper[4813]: I1206 16:50:36.313871 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b4wzv" Dec 06 16:50:36 crc kubenswrapper[4813]: I1206 16:50:36.895203 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b4wzv"] Dec 06 16:50:36 crc kubenswrapper[4813]: I1206 16:50:36.927209 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b4wzv" event={"ID":"9e6f68b4-913b-4d69-9c46-3e3e5f7634e1","Type":"ContainerStarted","Data":"084f204873f833a8c88e0a0fd7196b8d365edc4019d25fd40e7a9ffe90b8f88b"} Dec 06 16:50:37 crc kubenswrapper[4813]: I1206 16:50:37.936414 4813 generic.go:334] "Generic (PLEG): container finished" podID="9e6f68b4-913b-4d69-9c46-3e3e5f7634e1" containerID="72cca61a06fb6341728857baeb85bf10028bd6d101412d18087401ace1e5fd58" exitCode=0 Dec 06 16:50:37 crc kubenswrapper[4813]: I1206 16:50:37.936684 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b4wzv" event={"ID":"9e6f68b4-913b-4d69-9c46-3e3e5f7634e1","Type":"ContainerDied","Data":"72cca61a06fb6341728857baeb85bf10028bd6d101412d18087401ace1e5fd58"} Dec 06 16:50:37 crc kubenswrapper[4813]: I1206 16:50:37.938408 4813 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 16:50:38 crc kubenswrapper[4813]: I1206 16:50:38.948916 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b4wzv" event={"ID":"9e6f68b4-913b-4d69-9c46-3e3e5f7634e1","Type":"ContainerStarted","Data":"70d51cf73ae4fb8d3817473768d31a92c56c3fa388b1d7bbf73f3a94e1d7f061"} Dec 06 16:50:39 crc kubenswrapper[4813]: I1206 16:50:39.973411 4813 generic.go:334] "Generic (PLEG): container finished" podID="9e6f68b4-913b-4d69-9c46-3e3e5f7634e1" containerID="70d51cf73ae4fb8d3817473768d31a92c56c3fa388b1d7bbf73f3a94e1d7f061" exitCode=0 Dec 06 16:50:39 crc kubenswrapper[4813]: I1206 16:50:39.973475 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b4wzv" event={"ID":"9e6f68b4-913b-4d69-9c46-3e3e5f7634e1","Type":"ContainerDied","Data":"70d51cf73ae4fb8d3817473768d31a92c56c3fa388b1d7bbf73f3a94e1d7f061"} Dec 06 16:50:40 crc kubenswrapper[4813]: I1206 16:50:40.982400 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b4wzv" event={"ID":"9e6f68b4-913b-4d69-9c46-3e3e5f7634e1","Type":"ContainerStarted","Data":"1e7826d63b27b5c383df015b3895dd83b877bde3dc1255adac968ac55cad5890"} Dec 06 16:50:41 crc kubenswrapper[4813]: I1206 16:50:41.481018 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-mllt2_935ac385-32ee-46c5-bffb-69b89b4af6a9/control-plane-machine-set-operator/0.log" Dec 06 16:50:41 crc kubenswrapper[4813]: I1206 16:50:41.599771 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-dq6hp_0db3a3aa-f872-43c3-ade2-58c55a448fe3/kube-rbac-proxy/0.log" Dec 06 16:50:41 crc kubenswrapper[4813]: I1206 16:50:41.699620 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-dq6hp_0db3a3aa-f872-43c3-ade2-58c55a448fe3/machine-api-operator/0.log" Dec 06 16:50:42 crc kubenswrapper[4813]: I1206 16:50:42.487224 4813 scope.go:117] "RemoveContainer" containerID="969c7429b9faa2222909b4bad4ef827351b31f2cde6d6ab68e5af156f1ab0dfc" Dec 06 16:50:42 crc kubenswrapper[4813]: E1206 16:50:42.487580 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:50:46 crc kubenswrapper[4813]: I1206 16:50:46.315156 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-b4wzv" Dec 06 16:50:46 crc kubenswrapper[4813]: I1206 16:50:46.315598 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-b4wzv" Dec 06 16:50:46 crc kubenswrapper[4813]: I1206 16:50:46.386199 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-b4wzv" Dec 06 16:50:46 crc kubenswrapper[4813]: I1206 16:50:46.411951 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-b4wzv" podStartSLOduration=8.936715128 podStartE2EDuration="11.411933392s" podCreationTimestamp="2025-12-06 16:50:35 +0000 UTC" firstStartedPulling="2025-12-06 16:50:37.938111633 +0000 UTC m=+3877.828991219" lastFinishedPulling="2025-12-06 16:50:40.413329917 +0000 UTC m=+3880.304209483" observedRunningTime="2025-12-06 16:50:41.004318306 +0000 UTC m=+3880.895197882" watchObservedRunningTime="2025-12-06 16:50:46.411933392 +0000 UTC m=+3886.302812968" Dec 06 16:50:47 crc kubenswrapper[4813]: I1206 16:50:47.110659 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-b4wzv" Dec 06 16:50:47 crc kubenswrapper[4813]: I1206 16:50:47.189505 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b4wzv"] Dec 06 16:50:49 crc kubenswrapper[4813]: I1206 16:50:49.060049 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-b4wzv" podUID="9e6f68b4-913b-4d69-9c46-3e3e5f7634e1" containerName="registry-server" containerID="cri-o://1e7826d63b27b5c383df015b3895dd83b877bde3dc1255adac968ac55cad5890" gracePeriod=2 Dec 06 16:50:49 crc kubenswrapper[4813]: I1206 16:50:49.526428 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b4wzv" Dec 06 16:50:49 crc kubenswrapper[4813]: I1206 16:50:49.711173 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e6f68b4-913b-4d69-9c46-3e3e5f7634e1-catalog-content\") pod \"9e6f68b4-913b-4d69-9c46-3e3e5f7634e1\" (UID: \"9e6f68b4-913b-4d69-9c46-3e3e5f7634e1\") " Dec 06 16:50:49 crc kubenswrapper[4813]: I1206 16:50:49.711289 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtjsx\" (UniqueName: \"kubernetes.io/projected/9e6f68b4-913b-4d69-9c46-3e3e5f7634e1-kube-api-access-mtjsx\") pod \"9e6f68b4-913b-4d69-9c46-3e3e5f7634e1\" (UID: \"9e6f68b4-913b-4d69-9c46-3e3e5f7634e1\") " Dec 06 16:50:49 crc kubenswrapper[4813]: I1206 16:50:49.711314 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e6f68b4-913b-4d69-9c46-3e3e5f7634e1-utilities\") pod \"9e6f68b4-913b-4d69-9c46-3e3e5f7634e1\" (UID: \"9e6f68b4-913b-4d69-9c46-3e3e5f7634e1\") " Dec 06 16:50:49 crc kubenswrapper[4813]: I1206 16:50:49.712374 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e6f68b4-913b-4d69-9c46-3e3e5f7634e1-utilities" (OuterVolumeSpecName: "utilities") pod "9e6f68b4-913b-4d69-9c46-3e3e5f7634e1" (UID: "9e6f68b4-913b-4d69-9c46-3e3e5f7634e1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:50:49 crc kubenswrapper[4813]: I1206 16:50:49.716954 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e6f68b4-913b-4d69-9c46-3e3e5f7634e1-kube-api-access-mtjsx" (OuterVolumeSpecName: "kube-api-access-mtjsx") pod "9e6f68b4-913b-4d69-9c46-3e3e5f7634e1" (UID: "9e6f68b4-913b-4d69-9c46-3e3e5f7634e1"). InnerVolumeSpecName "kube-api-access-mtjsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:50:49 crc kubenswrapper[4813]: I1206 16:50:49.758187 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e6f68b4-913b-4d69-9c46-3e3e5f7634e1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9e6f68b4-913b-4d69-9c46-3e3e5f7634e1" (UID: "9e6f68b4-913b-4d69-9c46-3e3e5f7634e1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:50:49 crc kubenswrapper[4813]: I1206 16:50:49.814229 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtjsx\" (UniqueName: \"kubernetes.io/projected/9e6f68b4-913b-4d69-9c46-3e3e5f7634e1-kube-api-access-mtjsx\") on node \"crc\" DevicePath \"\"" Dec 06 16:50:49 crc kubenswrapper[4813]: I1206 16:50:49.814274 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e6f68b4-913b-4d69-9c46-3e3e5f7634e1-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 16:50:49 crc kubenswrapper[4813]: I1206 16:50:49.814286 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e6f68b4-913b-4d69-9c46-3e3e5f7634e1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 16:50:50 crc kubenswrapper[4813]: I1206 16:50:50.075766 4813 generic.go:334] "Generic (PLEG): container finished" podID="9e6f68b4-913b-4d69-9c46-3e3e5f7634e1" containerID="1e7826d63b27b5c383df015b3895dd83b877bde3dc1255adac968ac55cad5890" exitCode=0 Dec 06 16:50:50 crc kubenswrapper[4813]: I1206 16:50:50.075806 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b4wzv" event={"ID":"9e6f68b4-913b-4d69-9c46-3e3e5f7634e1","Type":"ContainerDied","Data":"1e7826d63b27b5c383df015b3895dd83b877bde3dc1255adac968ac55cad5890"} Dec 06 16:50:50 crc kubenswrapper[4813]: I1206 16:50:50.075833 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b4wzv" event={"ID":"9e6f68b4-913b-4d69-9c46-3e3e5f7634e1","Type":"ContainerDied","Data":"084f204873f833a8c88e0a0fd7196b8d365edc4019d25fd40e7a9ffe90b8f88b"} Dec 06 16:50:50 crc kubenswrapper[4813]: I1206 16:50:50.075839 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b4wzv" Dec 06 16:50:50 crc kubenswrapper[4813]: I1206 16:50:50.075848 4813 scope.go:117] "RemoveContainer" containerID="1e7826d63b27b5c383df015b3895dd83b877bde3dc1255adac968ac55cad5890" Dec 06 16:50:50 crc kubenswrapper[4813]: I1206 16:50:50.102549 4813 scope.go:117] "RemoveContainer" containerID="70d51cf73ae4fb8d3817473768d31a92c56c3fa388b1d7bbf73f3a94e1d7f061" Dec 06 16:50:50 crc kubenswrapper[4813]: I1206 16:50:50.126819 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b4wzv"] Dec 06 16:50:50 crc kubenswrapper[4813]: I1206 16:50:50.127704 4813 scope.go:117] "RemoveContainer" containerID="72cca61a06fb6341728857baeb85bf10028bd6d101412d18087401ace1e5fd58" Dec 06 16:50:50 crc kubenswrapper[4813]: I1206 16:50:50.136995 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-b4wzv"] Dec 06 16:50:50 crc kubenswrapper[4813]: I1206 16:50:50.184193 4813 scope.go:117] "RemoveContainer" containerID="1e7826d63b27b5c383df015b3895dd83b877bde3dc1255adac968ac55cad5890" Dec 06 16:50:50 crc kubenswrapper[4813]: E1206 16:50:50.184738 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e7826d63b27b5c383df015b3895dd83b877bde3dc1255adac968ac55cad5890\": container with ID starting with 1e7826d63b27b5c383df015b3895dd83b877bde3dc1255adac968ac55cad5890 not found: ID does not exist" containerID="1e7826d63b27b5c383df015b3895dd83b877bde3dc1255adac968ac55cad5890" Dec 06 16:50:50 crc kubenswrapper[4813]: I1206 16:50:50.184775 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e7826d63b27b5c383df015b3895dd83b877bde3dc1255adac968ac55cad5890"} err="failed to get container status \"1e7826d63b27b5c383df015b3895dd83b877bde3dc1255adac968ac55cad5890\": rpc error: code = NotFound desc = could not find container \"1e7826d63b27b5c383df015b3895dd83b877bde3dc1255adac968ac55cad5890\": container with ID starting with 1e7826d63b27b5c383df015b3895dd83b877bde3dc1255adac968ac55cad5890 not found: ID does not exist" Dec 06 16:50:50 crc kubenswrapper[4813]: I1206 16:50:50.184799 4813 scope.go:117] "RemoveContainer" containerID="70d51cf73ae4fb8d3817473768d31a92c56c3fa388b1d7bbf73f3a94e1d7f061" Dec 06 16:50:50 crc kubenswrapper[4813]: E1206 16:50:50.185120 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70d51cf73ae4fb8d3817473768d31a92c56c3fa388b1d7bbf73f3a94e1d7f061\": container with ID starting with 70d51cf73ae4fb8d3817473768d31a92c56c3fa388b1d7bbf73f3a94e1d7f061 not found: ID does not exist" containerID="70d51cf73ae4fb8d3817473768d31a92c56c3fa388b1d7bbf73f3a94e1d7f061" Dec 06 16:50:50 crc kubenswrapper[4813]: I1206 16:50:50.185144 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70d51cf73ae4fb8d3817473768d31a92c56c3fa388b1d7bbf73f3a94e1d7f061"} err="failed to get container status \"70d51cf73ae4fb8d3817473768d31a92c56c3fa388b1d7bbf73f3a94e1d7f061\": rpc error: code = NotFound desc = could not find container \"70d51cf73ae4fb8d3817473768d31a92c56c3fa388b1d7bbf73f3a94e1d7f061\": container with ID starting with 70d51cf73ae4fb8d3817473768d31a92c56c3fa388b1d7bbf73f3a94e1d7f061 not found: ID does not exist" Dec 06 16:50:50 crc kubenswrapper[4813]: I1206 16:50:50.185158 4813 scope.go:117] "RemoveContainer" containerID="72cca61a06fb6341728857baeb85bf10028bd6d101412d18087401ace1e5fd58" Dec 06 16:50:50 crc kubenswrapper[4813]: E1206 16:50:50.185449 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72cca61a06fb6341728857baeb85bf10028bd6d101412d18087401ace1e5fd58\": container with ID starting with 72cca61a06fb6341728857baeb85bf10028bd6d101412d18087401ace1e5fd58 not found: ID does not exist" containerID="72cca61a06fb6341728857baeb85bf10028bd6d101412d18087401ace1e5fd58" Dec 06 16:50:50 crc kubenswrapper[4813]: I1206 16:50:50.185472 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72cca61a06fb6341728857baeb85bf10028bd6d101412d18087401ace1e5fd58"} err="failed to get container status \"72cca61a06fb6341728857baeb85bf10028bd6d101412d18087401ace1e5fd58\": rpc error: code = NotFound desc = could not find container \"72cca61a06fb6341728857baeb85bf10028bd6d101412d18087401ace1e5fd58\": container with ID starting with 72cca61a06fb6341728857baeb85bf10028bd6d101412d18087401ace1e5fd58 not found: ID does not exist" Dec 06 16:50:50 crc kubenswrapper[4813]: I1206 16:50:50.503813 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e6f68b4-913b-4d69-9c46-3e3e5f7634e1" path="/var/lib/kubelet/pods/9e6f68b4-913b-4d69-9c46-3e3e5f7634e1/volumes" Dec 06 16:50:56 crc kubenswrapper[4813]: I1206 16:50:56.487334 4813 scope.go:117] "RemoveContainer" containerID="969c7429b9faa2222909b4bad4ef827351b31f2cde6d6ab68e5af156f1ab0dfc" Dec 06 16:50:56 crc kubenswrapper[4813]: E1206 16:50:56.488025 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:50:57 crc kubenswrapper[4813]: I1206 16:50:57.336945 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-gqwzr_9a87a677-1122-46a6-aae7-820387f122d5/cert-manager-controller/0.log" Dec 06 16:50:57 crc kubenswrapper[4813]: I1206 16:50:57.410101 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-6jx97_93a0448f-03c2-4a73-8ebc-b6bc1a5a3cd3/cert-manager-cainjector/0.log" Dec 06 16:50:57 crc kubenswrapper[4813]: I1206 16:50:57.490631 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-4m6ft_b5fd38e1-0a32-486d-839d-e3cce8c150fe/cert-manager-webhook/0.log" Dec 06 16:51:10 crc kubenswrapper[4813]: I1206 16:51:10.704501 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-r7rqn_2e477e7f-f00a-40a3-b26a-d49ef1f7f41d/nmstate-console-plugin/0.log" Dec 06 16:51:10 crc kubenswrapper[4813]: I1206 16:51:10.925542 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-f4x5w_db7dde13-c533-49ae-bae9-7239ad2b8db6/nmstate-metrics/0.log" Dec 06 16:51:10 crc kubenswrapper[4813]: I1206 16:51:10.953370 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-zw99x_4f31e85e-1c68-4627-b047-72de6b05d490/nmstate-handler/0.log" Dec 06 16:51:10 crc kubenswrapper[4813]: I1206 16:51:10.956722 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-f4x5w_db7dde13-c533-49ae-bae9-7239ad2b8db6/kube-rbac-proxy/0.log" Dec 06 16:51:11 crc kubenswrapper[4813]: I1206 16:51:11.161825 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-bd2wf_215ac015-9856-4c78-ab4d-0035ee29b678/nmstate-operator/0.log" Dec 06 16:51:11 crc kubenswrapper[4813]: I1206 16:51:11.165405 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-vc5f4_3b585cd5-6045-46b1-86cf-03536009e643/nmstate-webhook/0.log" Dec 06 16:51:11 crc kubenswrapper[4813]: I1206 16:51:11.487311 4813 scope.go:117] "RemoveContainer" containerID="969c7429b9faa2222909b4bad4ef827351b31f2cde6d6ab68e5af156f1ab0dfc" Dec 06 16:51:11 crc kubenswrapper[4813]: E1206 16:51:11.487537 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:51:25 crc kubenswrapper[4813]: I1206 16:51:25.486871 4813 scope.go:117] "RemoveContainer" containerID="969c7429b9faa2222909b4bad4ef827351b31f2cde6d6ab68e5af156f1ab0dfc" Dec 06 16:51:25 crc kubenswrapper[4813]: E1206 16:51:25.487671 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:51:29 crc kubenswrapper[4813]: I1206 16:51:29.449017 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-mfvl4_a8a9650e-6412-4d98-bdf5-6b57f014eb64/kube-rbac-proxy/0.log" Dec 06 16:51:29 crc kubenswrapper[4813]: I1206 16:51:29.456956 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-mfvl4_a8a9650e-6412-4d98-bdf5-6b57f014eb64/controller/0.log" Dec 06 16:51:30 crc kubenswrapper[4813]: I1206 16:51:30.017737 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/cp-frr-files/0.log" Dec 06 16:51:30 crc kubenswrapper[4813]: I1206 16:51:30.281550 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/cp-reloader/0.log" Dec 06 16:51:30 crc kubenswrapper[4813]: I1206 16:51:30.354877 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/cp-metrics/0.log" Dec 06 16:51:30 crc kubenswrapper[4813]: I1206 16:51:30.367941 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/cp-frr-files/0.log" Dec 06 16:51:30 crc kubenswrapper[4813]: I1206 16:51:30.416188 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/cp-reloader/0.log" Dec 06 16:51:30 crc kubenswrapper[4813]: I1206 16:51:30.666141 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/cp-reloader/0.log" Dec 06 16:51:30 crc kubenswrapper[4813]: I1206 16:51:30.676021 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/cp-metrics/0.log" Dec 06 16:51:30 crc kubenswrapper[4813]: I1206 16:51:30.699492 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/cp-frr-files/0.log" Dec 06 16:51:30 crc kubenswrapper[4813]: I1206 16:51:30.740142 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/cp-metrics/0.log" Dec 06 16:51:30 crc kubenswrapper[4813]: I1206 16:51:30.941687 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/cp-reloader/0.log" Dec 06 16:51:30 crc kubenswrapper[4813]: I1206 16:51:30.950272 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/cp-frr-files/0.log" Dec 06 16:51:31 crc kubenswrapper[4813]: I1206 16:51:31.000235 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/controller/0.log" Dec 06 16:51:31 crc kubenswrapper[4813]: I1206 16:51:31.007777 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/cp-metrics/0.log" Dec 06 16:51:31 crc kubenswrapper[4813]: I1206 16:51:31.166829 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/frr-metrics/0.log" Dec 06 16:51:31 crc kubenswrapper[4813]: I1206 16:51:31.267665 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/kube-rbac-proxy/0.log" Dec 06 16:51:31 crc kubenswrapper[4813]: I1206 16:51:31.268930 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/kube-rbac-proxy-frr/0.log" Dec 06 16:51:31 crc kubenswrapper[4813]: I1206 16:51:31.484476 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/reloader/0.log" Dec 06 16:51:31 crc kubenswrapper[4813]: I1206 16:51:31.589393 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-wsvkk_a5c26c26-14a5-406c-882b-2950fd2855ed/frr-k8s-webhook-server/0.log" Dec 06 16:51:31 crc kubenswrapper[4813]: I1206 16:51:31.873817 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6977db5dfd-44msc_e7537632-d133-4c59-a167-e59127ef6fab/manager/0.log" Dec 06 16:51:31 crc kubenswrapper[4813]: I1206 16:51:31.997073 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-65d646cc59-hk6wv_2fcdfd68-9b08-423f-80ec-53032016fdfe/webhook-server/0.log" Dec 06 16:51:32 crc kubenswrapper[4813]: I1206 16:51:32.160001 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tdbkn_74e697b0-ef6a-45c2-8671-c67725e42ab6/frr/0.log" Dec 06 16:51:32 crc kubenswrapper[4813]: I1206 16:51:32.275374 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-ccll7_45152d05-2dee-4bba-bc56-2e378d0edb9a/kube-rbac-proxy/0.log" Dec 06 16:51:32 crc kubenswrapper[4813]: I1206 16:51:32.591122 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-ccll7_45152d05-2dee-4bba-bc56-2e378d0edb9a/speaker/0.log" Dec 06 16:51:39 crc kubenswrapper[4813]: I1206 16:51:39.515070 4813 scope.go:117] "RemoveContainer" containerID="969c7429b9faa2222909b4bad4ef827351b31f2cde6d6ab68e5af156f1ab0dfc" Dec 06 16:51:39 crc kubenswrapper[4813]: E1206 16:51:39.516293 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t5xp8_openshift-machine-config-operator(d88e8bae-c055-4c55-b548-f621ff96de06)\"" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" Dec 06 16:51:47 crc kubenswrapper[4813]: I1206 16:51:47.183781 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx_4cc38736-85b3-4e2d-87fe-604eebf52317/util/0.log" Dec 06 16:51:47 crc kubenswrapper[4813]: I1206 16:51:47.363857 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx_4cc38736-85b3-4e2d-87fe-604eebf52317/pull/0.log" Dec 06 16:51:47 crc kubenswrapper[4813]: I1206 16:51:47.376854 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx_4cc38736-85b3-4e2d-87fe-604eebf52317/util/0.log" Dec 06 16:51:47 crc kubenswrapper[4813]: I1206 16:51:47.479301 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx_4cc38736-85b3-4e2d-87fe-604eebf52317/pull/0.log" Dec 06 16:51:47 crc kubenswrapper[4813]: I1206 16:51:47.645601 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx_4cc38736-85b3-4e2d-87fe-604eebf52317/pull/0.log" Dec 06 16:51:47 crc kubenswrapper[4813]: I1206 16:51:47.648471 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx_4cc38736-85b3-4e2d-87fe-604eebf52317/util/0.log" Dec 06 16:51:47 crc kubenswrapper[4813]: I1206 16:51:47.718579 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjrwwx_4cc38736-85b3-4e2d-87fe-604eebf52317/extract/0.log" Dec 06 16:51:48 crc kubenswrapper[4813]: I1206 16:51:48.162667 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5_6f21e1e0-2a67-4f45-93f4-e84a5bba3007/util/0.log" Dec 06 16:51:48 crc kubenswrapper[4813]: I1206 16:51:48.313431 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5_6f21e1e0-2a67-4f45-93f4-e84a5bba3007/util/0.log" Dec 06 16:51:48 crc kubenswrapper[4813]: I1206 16:51:48.321289 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5_6f21e1e0-2a67-4f45-93f4-e84a5bba3007/pull/0.log" Dec 06 16:51:48 crc kubenswrapper[4813]: I1206 16:51:48.348817 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5_6f21e1e0-2a67-4f45-93f4-e84a5bba3007/pull/0.log" Dec 06 16:51:48 crc kubenswrapper[4813]: I1206 16:51:48.554414 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5_6f21e1e0-2a67-4f45-93f4-e84a5bba3007/util/0.log" Dec 06 16:51:48 crc kubenswrapper[4813]: I1206 16:51:48.568306 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5_6f21e1e0-2a67-4f45-93f4-e84a5bba3007/pull/0.log" Dec 06 16:51:48 crc kubenswrapper[4813]: I1206 16:51:48.656205 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cx5k5_6f21e1e0-2a67-4f45-93f4-e84a5bba3007/extract/0.log" Dec 06 16:51:48 crc kubenswrapper[4813]: I1206 16:51:48.825079 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ws47p_abeb6c67-8702-4211-9aab-7cdf25c0d078/extract-utilities/0.log" Dec 06 16:51:49 crc kubenswrapper[4813]: I1206 16:51:49.028431 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ws47p_abeb6c67-8702-4211-9aab-7cdf25c0d078/extract-content/0.log" Dec 06 16:51:49 crc kubenswrapper[4813]: I1206 16:51:49.068847 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ws47p_abeb6c67-8702-4211-9aab-7cdf25c0d078/extract-utilities/0.log" Dec 06 16:51:49 crc kubenswrapper[4813]: I1206 16:51:49.148353 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ws47p_abeb6c67-8702-4211-9aab-7cdf25c0d078/extract-content/0.log" Dec 06 16:51:49 crc kubenswrapper[4813]: I1206 16:51:49.372316 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ws47p_abeb6c67-8702-4211-9aab-7cdf25c0d078/extract-content/0.log" Dec 06 16:51:49 crc kubenswrapper[4813]: I1206 16:51:49.373116 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ws47p_abeb6c67-8702-4211-9aab-7cdf25c0d078/extract-utilities/0.log" Dec 06 16:51:49 crc kubenswrapper[4813]: I1206 16:51:49.699942 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-crqpg_4b8fe11e-abdb-42d7-9d88-7d463a808b4b/extract-utilities/0.log" Dec 06 16:51:49 crc kubenswrapper[4813]: I1206 16:51:49.767097 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ws47p_abeb6c67-8702-4211-9aab-7cdf25c0d078/registry-server/0.log" Dec 06 16:51:49 crc kubenswrapper[4813]: I1206 16:51:49.914288 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-crqpg_4b8fe11e-abdb-42d7-9d88-7d463a808b4b/extract-content/0.log" Dec 06 16:51:49 crc kubenswrapper[4813]: I1206 16:51:49.944023 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-crqpg_4b8fe11e-abdb-42d7-9d88-7d463a808b4b/extract-content/0.log" Dec 06 16:51:49 crc kubenswrapper[4813]: I1206 16:51:49.972420 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-crqpg_4b8fe11e-abdb-42d7-9d88-7d463a808b4b/extract-utilities/0.log" Dec 06 16:51:50 crc kubenswrapper[4813]: I1206 16:51:50.213774 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-crqpg_4b8fe11e-abdb-42d7-9d88-7d463a808b4b/extract-utilities/0.log" Dec 06 16:51:50 crc kubenswrapper[4813]: I1206 16:51:50.221704 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-crqpg_4b8fe11e-abdb-42d7-9d88-7d463a808b4b/extract-content/0.log" Dec 06 16:51:50 crc kubenswrapper[4813]: I1206 16:51:50.468949 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-crqpg_4b8fe11e-abdb-42d7-9d88-7d463a808b4b/registry-server/0.log" Dec 06 16:51:50 crc kubenswrapper[4813]: I1206 16:51:50.548745 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-v6fmp_2660f707-b9cc-4c4f-b42a-9aa6699ef326/marketplace-operator/0.log" Dec 06 16:51:50 crc kubenswrapper[4813]: I1206 16:51:50.601902 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xwq7l_ca46ce83-a423-478a-9d44-d45a0ef28ef3/extract-utilities/0.log" Dec 06 16:51:50 crc kubenswrapper[4813]: I1206 16:51:50.712242 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xwq7l_ca46ce83-a423-478a-9d44-d45a0ef28ef3/extract-utilities/0.log" Dec 06 16:51:50 crc kubenswrapper[4813]: I1206 16:51:50.740449 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xwq7l_ca46ce83-a423-478a-9d44-d45a0ef28ef3/extract-content/0.log" Dec 06 16:51:50 crc kubenswrapper[4813]: I1206 16:51:50.832070 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xwq7l_ca46ce83-a423-478a-9d44-d45a0ef28ef3/extract-content/0.log" Dec 06 16:51:50 crc kubenswrapper[4813]: I1206 16:51:50.948361 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xwq7l_ca46ce83-a423-478a-9d44-d45a0ef28ef3/extract-utilities/0.log" Dec 06 16:51:50 crc kubenswrapper[4813]: I1206 16:51:50.982909 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xwq7l_ca46ce83-a423-478a-9d44-d45a0ef28ef3/extract-content/0.log" Dec 06 16:51:51 crc kubenswrapper[4813]: I1206 16:51:51.280964 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xwq7l_ca46ce83-a423-478a-9d44-d45a0ef28ef3/registry-server/0.log" Dec 06 16:51:51 crc kubenswrapper[4813]: I1206 16:51:51.306275 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-28zj9_46d12276-3ced-4cbe-a060-704f83fcf9a8/extract-utilities/0.log" Dec 06 16:51:51 crc kubenswrapper[4813]: I1206 16:51:51.479410 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-28zj9_46d12276-3ced-4cbe-a060-704f83fcf9a8/extract-utilities/0.log" Dec 06 16:51:51 crc kubenswrapper[4813]: I1206 16:51:51.490768 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-28zj9_46d12276-3ced-4cbe-a060-704f83fcf9a8/extract-content/0.log" Dec 06 16:51:51 crc kubenswrapper[4813]: I1206 16:51:51.502735 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-28zj9_46d12276-3ced-4cbe-a060-704f83fcf9a8/extract-content/0.log" Dec 06 16:51:51 crc kubenswrapper[4813]: I1206 16:51:51.743503 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-28zj9_46d12276-3ced-4cbe-a060-704f83fcf9a8/extract-utilities/0.log" Dec 06 16:51:51 crc kubenswrapper[4813]: I1206 16:51:51.758481 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-28zj9_46d12276-3ced-4cbe-a060-704f83fcf9a8/extract-content/0.log" Dec 06 16:51:52 crc kubenswrapper[4813]: I1206 16:51:52.120553 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-28zj9_46d12276-3ced-4cbe-a060-704f83fcf9a8/registry-server/0.log" Dec 06 16:51:54 crc kubenswrapper[4813]: I1206 16:51:54.487339 4813 scope.go:117] "RemoveContainer" containerID="969c7429b9faa2222909b4bad4ef827351b31f2cde6d6ab68e5af156f1ab0dfc" Dec 06 16:51:55 crc kubenswrapper[4813]: I1206 16:51:55.674006 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerStarted","Data":"fcb495a2496cefa1a08b361825efb708538fc1c2d99f9878beb2279f9088eca6"} Dec 06 16:52:23 crc kubenswrapper[4813]: E1206 16:52:23.608889 4813 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.9:54464->38.102.83.9:38901: write tcp 38.102.83.9:54464->38.102.83.9:38901: write: connection reset by peer Dec 06 16:52:23 crc kubenswrapper[4813]: E1206 16:52:23.652546 4813 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.9:54538->38.102.83.9:38901: write tcp 38.102.83.9:54538->38.102.83.9:38901: write: connection reset by peer Dec 06 16:53:36 crc kubenswrapper[4813]: I1206 16:53:36.713841 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8nxxf"] Dec 06 16:53:36 crc kubenswrapper[4813]: E1206 16:53:36.715851 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e6f68b4-913b-4d69-9c46-3e3e5f7634e1" containerName="extract-utilities" Dec 06 16:53:36 crc kubenswrapper[4813]: I1206 16:53:36.716055 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e6f68b4-913b-4d69-9c46-3e3e5f7634e1" containerName="extract-utilities" Dec 06 16:53:36 crc kubenswrapper[4813]: E1206 16:53:36.716986 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e6f68b4-913b-4d69-9c46-3e3e5f7634e1" containerName="registry-server" Dec 06 16:53:36 crc kubenswrapper[4813]: I1206 16:53:36.717063 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e6f68b4-913b-4d69-9c46-3e3e5f7634e1" containerName="registry-server" Dec 06 16:53:36 crc kubenswrapper[4813]: E1206 16:53:36.717128 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e6f68b4-913b-4d69-9c46-3e3e5f7634e1" containerName="extract-content" Dec 06 16:53:36 crc kubenswrapper[4813]: I1206 16:53:36.717135 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e6f68b4-913b-4d69-9c46-3e3e5f7634e1" containerName="extract-content" Dec 06 16:53:36 crc kubenswrapper[4813]: I1206 16:53:36.717474 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e6f68b4-913b-4d69-9c46-3e3e5f7634e1" containerName="registry-server" Dec 06 16:53:36 crc kubenswrapper[4813]: I1206 16:53:36.718979 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8nxxf" Dec 06 16:53:36 crc kubenswrapper[4813]: I1206 16:53:36.735334 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8nxxf"] Dec 06 16:53:36 crc kubenswrapper[4813]: I1206 16:53:36.911091 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e5505e6-cb46-4c1a-81d6-3f4d970291e7-utilities\") pod \"community-operators-8nxxf\" (UID: \"8e5505e6-cb46-4c1a-81d6-3f4d970291e7\") " pod="openshift-marketplace/community-operators-8nxxf" Dec 06 16:53:36 crc kubenswrapper[4813]: I1206 16:53:36.911198 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e5505e6-cb46-4c1a-81d6-3f4d970291e7-catalog-content\") pod \"community-operators-8nxxf\" (UID: \"8e5505e6-cb46-4c1a-81d6-3f4d970291e7\") " pod="openshift-marketplace/community-operators-8nxxf" Dec 06 16:53:36 crc kubenswrapper[4813]: I1206 16:53:36.911287 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llcth\" (UniqueName: \"kubernetes.io/projected/8e5505e6-cb46-4c1a-81d6-3f4d970291e7-kube-api-access-llcth\") pod \"community-operators-8nxxf\" (UID: \"8e5505e6-cb46-4c1a-81d6-3f4d970291e7\") " pod="openshift-marketplace/community-operators-8nxxf" Dec 06 16:53:37 crc kubenswrapper[4813]: I1206 16:53:37.012569 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e5505e6-cb46-4c1a-81d6-3f4d970291e7-catalog-content\") pod \"community-operators-8nxxf\" (UID: \"8e5505e6-cb46-4c1a-81d6-3f4d970291e7\") " pod="openshift-marketplace/community-operators-8nxxf" Dec 06 16:53:37 crc kubenswrapper[4813]: I1206 16:53:37.012677 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llcth\" (UniqueName: \"kubernetes.io/projected/8e5505e6-cb46-4c1a-81d6-3f4d970291e7-kube-api-access-llcth\") pod \"community-operators-8nxxf\" (UID: \"8e5505e6-cb46-4c1a-81d6-3f4d970291e7\") " pod="openshift-marketplace/community-operators-8nxxf" Dec 06 16:53:37 crc kubenswrapper[4813]: I1206 16:53:37.012709 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e5505e6-cb46-4c1a-81d6-3f4d970291e7-utilities\") pod \"community-operators-8nxxf\" (UID: \"8e5505e6-cb46-4c1a-81d6-3f4d970291e7\") " pod="openshift-marketplace/community-operators-8nxxf" Dec 06 16:53:37 crc kubenswrapper[4813]: I1206 16:53:37.013180 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e5505e6-cb46-4c1a-81d6-3f4d970291e7-catalog-content\") pod \"community-operators-8nxxf\" (UID: \"8e5505e6-cb46-4c1a-81d6-3f4d970291e7\") " pod="openshift-marketplace/community-operators-8nxxf" Dec 06 16:53:37 crc kubenswrapper[4813]: I1206 16:53:37.013192 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e5505e6-cb46-4c1a-81d6-3f4d970291e7-utilities\") pod \"community-operators-8nxxf\" (UID: \"8e5505e6-cb46-4c1a-81d6-3f4d970291e7\") " pod="openshift-marketplace/community-operators-8nxxf" Dec 06 16:53:37 crc kubenswrapper[4813]: I1206 16:53:37.035347 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llcth\" (UniqueName: \"kubernetes.io/projected/8e5505e6-cb46-4c1a-81d6-3f4d970291e7-kube-api-access-llcth\") pod \"community-operators-8nxxf\" (UID: \"8e5505e6-cb46-4c1a-81d6-3f4d970291e7\") " pod="openshift-marketplace/community-operators-8nxxf" Dec 06 16:53:37 crc kubenswrapper[4813]: I1206 16:53:37.050094 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8nxxf" Dec 06 16:53:37 crc kubenswrapper[4813]: I1206 16:53:37.573570 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8nxxf"] Dec 06 16:53:37 crc kubenswrapper[4813]: W1206 16:53:37.583864 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e5505e6_cb46_4c1a_81d6_3f4d970291e7.slice/crio-342404d4f49b70e065e56895bcbd93727af2bb6dddd054d2314c765f51d01bd5 WatchSource:0}: Error finding container 342404d4f49b70e065e56895bcbd93727af2bb6dddd054d2314c765f51d01bd5: Status 404 returned error can't find the container with id 342404d4f49b70e065e56895bcbd93727af2bb6dddd054d2314c765f51d01bd5 Dec 06 16:53:37 crc kubenswrapper[4813]: I1206 16:53:37.773641 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8nxxf" event={"ID":"8e5505e6-cb46-4c1a-81d6-3f4d970291e7","Type":"ContainerStarted","Data":"342404d4f49b70e065e56895bcbd93727af2bb6dddd054d2314c765f51d01bd5"} Dec 06 16:53:38 crc kubenswrapper[4813]: I1206 16:53:38.786090 4813 generic.go:334] "Generic (PLEG): container finished" podID="8e5505e6-cb46-4c1a-81d6-3f4d970291e7" containerID="b87b9060f94bb5e5d2af15a74f6ac48e12bb89c8c9cb8e37f4d1431950a367d8" exitCode=0 Dec 06 16:53:38 crc kubenswrapper[4813]: I1206 16:53:38.786209 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8nxxf" event={"ID":"8e5505e6-cb46-4c1a-81d6-3f4d970291e7","Type":"ContainerDied","Data":"b87b9060f94bb5e5d2af15a74f6ac48e12bb89c8c9cb8e37f4d1431950a367d8"} Dec 06 16:53:39 crc kubenswrapper[4813]: I1206 16:53:39.800380 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8nxxf" event={"ID":"8e5505e6-cb46-4c1a-81d6-3f4d970291e7","Type":"ContainerStarted","Data":"98c3a84bb3719670b27f639d1ea40727f5a8a516d9dc7b62648e43e0b2b56c6c"} Dec 06 16:53:40 crc kubenswrapper[4813]: I1206 16:53:40.810532 4813 generic.go:334] "Generic (PLEG): container finished" podID="8e5505e6-cb46-4c1a-81d6-3f4d970291e7" containerID="98c3a84bb3719670b27f639d1ea40727f5a8a516d9dc7b62648e43e0b2b56c6c" exitCode=0 Dec 06 16:53:40 crc kubenswrapper[4813]: I1206 16:53:40.810876 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8nxxf" event={"ID":"8e5505e6-cb46-4c1a-81d6-3f4d970291e7","Type":"ContainerDied","Data":"98c3a84bb3719670b27f639d1ea40727f5a8a516d9dc7b62648e43e0b2b56c6c"} Dec 06 16:53:41 crc kubenswrapper[4813]: I1206 16:53:41.825174 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8nxxf" event={"ID":"8e5505e6-cb46-4c1a-81d6-3f4d970291e7","Type":"ContainerStarted","Data":"8d21945e5f1a17d8ef7f062b987af5f440623d547bcaffa62b1c9fc794bf2694"} Dec 06 16:53:41 crc kubenswrapper[4813]: I1206 16:53:41.831706 4813 generic.go:334] "Generic (PLEG): container finished" podID="1a835870-e23a-446e-8c38-fe587e9cee6b" containerID="db9b299109c15116a10e92800ddbe7711c0fbc10c48090abc39a6e10048f6867" exitCode=0 Dec 06 16:53:41 crc kubenswrapper[4813]: I1206 16:53:41.831753 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jgn6l/must-gather-jb2rq" event={"ID":"1a835870-e23a-446e-8c38-fe587e9cee6b","Type":"ContainerDied","Data":"db9b299109c15116a10e92800ddbe7711c0fbc10c48090abc39a6e10048f6867"} Dec 06 16:53:41 crc kubenswrapper[4813]: I1206 16:53:41.832576 4813 scope.go:117] "RemoveContainer" containerID="db9b299109c15116a10e92800ddbe7711c0fbc10c48090abc39a6e10048f6867" Dec 06 16:53:41 crc kubenswrapper[4813]: I1206 16:53:41.847155 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8nxxf" podStartSLOduration=3.073066921 podStartE2EDuration="5.847129293s" podCreationTimestamp="2025-12-06 16:53:36 +0000 UTC" firstStartedPulling="2025-12-06 16:53:38.790157733 +0000 UTC m=+4058.681037349" lastFinishedPulling="2025-12-06 16:53:41.564220135 +0000 UTC m=+4061.455099721" observedRunningTime="2025-12-06 16:53:41.844056752 +0000 UTC m=+4061.734936338" watchObservedRunningTime="2025-12-06 16:53:41.847129293 +0000 UTC m=+4061.738008909" Dec 06 16:53:42 crc kubenswrapper[4813]: I1206 16:53:42.649354 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jgn6l_must-gather-jb2rq_1a835870-e23a-446e-8c38-fe587e9cee6b/gather/0.log" Dec 06 16:53:47 crc kubenswrapper[4813]: I1206 16:53:47.051324 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8nxxf" Dec 06 16:53:47 crc kubenswrapper[4813]: I1206 16:53:47.052142 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8nxxf" Dec 06 16:53:47 crc kubenswrapper[4813]: I1206 16:53:47.954380 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8nxxf" Dec 06 16:53:48 crc kubenswrapper[4813]: I1206 16:53:48.040964 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8nxxf" Dec 06 16:53:48 crc kubenswrapper[4813]: I1206 16:53:48.206683 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8nxxf"] Dec 06 16:53:49 crc kubenswrapper[4813]: I1206 16:53:49.918243 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8nxxf" podUID="8e5505e6-cb46-4c1a-81d6-3f4d970291e7" containerName="registry-server" containerID="cri-o://8d21945e5f1a17d8ef7f062b987af5f440623d547bcaffa62b1c9fc794bf2694" gracePeriod=2 Dec 06 16:53:50 crc kubenswrapper[4813]: I1206 16:53:50.476917 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8nxxf" Dec 06 16:53:50 crc kubenswrapper[4813]: I1206 16:53:50.654387 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e5505e6-cb46-4c1a-81d6-3f4d970291e7-catalog-content\") pod \"8e5505e6-cb46-4c1a-81d6-3f4d970291e7\" (UID: \"8e5505e6-cb46-4c1a-81d6-3f4d970291e7\") " Dec 06 16:53:50 crc kubenswrapper[4813]: I1206 16:53:50.661570 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llcth\" (UniqueName: \"kubernetes.io/projected/8e5505e6-cb46-4c1a-81d6-3f4d970291e7-kube-api-access-llcth\") pod \"8e5505e6-cb46-4c1a-81d6-3f4d970291e7\" (UID: \"8e5505e6-cb46-4c1a-81d6-3f4d970291e7\") " Dec 06 16:53:50 crc kubenswrapper[4813]: I1206 16:53:50.661628 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e5505e6-cb46-4c1a-81d6-3f4d970291e7-utilities\") pod \"8e5505e6-cb46-4c1a-81d6-3f4d970291e7\" (UID: \"8e5505e6-cb46-4c1a-81d6-3f4d970291e7\") " Dec 06 16:53:50 crc kubenswrapper[4813]: I1206 16:53:50.663464 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e5505e6-cb46-4c1a-81d6-3f4d970291e7-utilities" (OuterVolumeSpecName: "utilities") pod "8e5505e6-cb46-4c1a-81d6-3f4d970291e7" (UID: "8e5505e6-cb46-4c1a-81d6-3f4d970291e7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:53:50 crc kubenswrapper[4813]: I1206 16:53:50.671696 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e5505e6-cb46-4c1a-81d6-3f4d970291e7-kube-api-access-llcth" (OuterVolumeSpecName: "kube-api-access-llcth") pod "8e5505e6-cb46-4c1a-81d6-3f4d970291e7" (UID: "8e5505e6-cb46-4c1a-81d6-3f4d970291e7"). InnerVolumeSpecName "kube-api-access-llcth". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:53:50 crc kubenswrapper[4813]: I1206 16:53:50.723624 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e5505e6-cb46-4c1a-81d6-3f4d970291e7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8e5505e6-cb46-4c1a-81d6-3f4d970291e7" (UID: "8e5505e6-cb46-4c1a-81d6-3f4d970291e7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:53:50 crc kubenswrapper[4813]: I1206 16:53:50.765564 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e5505e6-cb46-4c1a-81d6-3f4d970291e7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 16:53:50 crc kubenswrapper[4813]: I1206 16:53:50.765592 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llcth\" (UniqueName: \"kubernetes.io/projected/8e5505e6-cb46-4c1a-81d6-3f4d970291e7-kube-api-access-llcth\") on node \"crc\" DevicePath \"\"" Dec 06 16:53:50 crc kubenswrapper[4813]: I1206 16:53:50.765605 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e5505e6-cb46-4c1a-81d6-3f4d970291e7-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 16:53:50 crc kubenswrapper[4813]: I1206 16:53:50.931019 4813 generic.go:334] "Generic (PLEG): container finished" podID="8e5505e6-cb46-4c1a-81d6-3f4d970291e7" containerID="8d21945e5f1a17d8ef7f062b987af5f440623d547bcaffa62b1c9fc794bf2694" exitCode=0 Dec 06 16:53:50 crc kubenswrapper[4813]: I1206 16:53:50.931091 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8nxxf" event={"ID":"8e5505e6-cb46-4c1a-81d6-3f4d970291e7","Type":"ContainerDied","Data":"8d21945e5f1a17d8ef7f062b987af5f440623d547bcaffa62b1c9fc794bf2694"} Dec 06 16:53:50 crc kubenswrapper[4813]: I1206 16:53:50.931124 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8nxxf" Dec 06 16:53:50 crc kubenswrapper[4813]: I1206 16:53:50.932416 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8nxxf" event={"ID":"8e5505e6-cb46-4c1a-81d6-3f4d970291e7","Type":"ContainerDied","Data":"342404d4f49b70e065e56895bcbd93727af2bb6dddd054d2314c765f51d01bd5"} Dec 06 16:53:50 crc kubenswrapper[4813]: I1206 16:53:50.932602 4813 scope.go:117] "RemoveContainer" containerID="8d21945e5f1a17d8ef7f062b987af5f440623d547bcaffa62b1c9fc794bf2694" Dec 06 16:53:50 crc kubenswrapper[4813]: I1206 16:53:50.974013 4813 scope.go:117] "RemoveContainer" containerID="98c3a84bb3719670b27f639d1ea40727f5a8a516d9dc7b62648e43e0b2b56c6c" Dec 06 16:53:50 crc kubenswrapper[4813]: I1206 16:53:50.989745 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8nxxf"] Dec 06 16:53:50 crc kubenswrapper[4813]: I1206 16:53:50.996821 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8nxxf"] Dec 06 16:53:51 crc kubenswrapper[4813]: I1206 16:53:51.001853 4813 scope.go:117] "RemoveContainer" containerID="b87b9060f94bb5e5d2af15a74f6ac48e12bb89c8c9cb8e37f4d1431950a367d8" Dec 06 16:53:51 crc kubenswrapper[4813]: I1206 16:53:51.037619 4813 scope.go:117] "RemoveContainer" containerID="8d21945e5f1a17d8ef7f062b987af5f440623d547bcaffa62b1c9fc794bf2694" Dec 06 16:53:51 crc kubenswrapper[4813]: E1206 16:53:51.038247 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d21945e5f1a17d8ef7f062b987af5f440623d547bcaffa62b1c9fc794bf2694\": container with ID starting with 8d21945e5f1a17d8ef7f062b987af5f440623d547bcaffa62b1c9fc794bf2694 not found: ID does not exist" containerID="8d21945e5f1a17d8ef7f062b987af5f440623d547bcaffa62b1c9fc794bf2694" Dec 06 16:53:51 crc kubenswrapper[4813]: I1206 16:53:51.038368 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d21945e5f1a17d8ef7f062b987af5f440623d547bcaffa62b1c9fc794bf2694"} err="failed to get container status \"8d21945e5f1a17d8ef7f062b987af5f440623d547bcaffa62b1c9fc794bf2694\": rpc error: code = NotFound desc = could not find container \"8d21945e5f1a17d8ef7f062b987af5f440623d547bcaffa62b1c9fc794bf2694\": container with ID starting with 8d21945e5f1a17d8ef7f062b987af5f440623d547bcaffa62b1c9fc794bf2694 not found: ID does not exist" Dec 06 16:53:51 crc kubenswrapper[4813]: I1206 16:53:51.038456 4813 scope.go:117] "RemoveContainer" containerID="98c3a84bb3719670b27f639d1ea40727f5a8a516d9dc7b62648e43e0b2b56c6c" Dec 06 16:53:51 crc kubenswrapper[4813]: E1206 16:53:51.038798 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98c3a84bb3719670b27f639d1ea40727f5a8a516d9dc7b62648e43e0b2b56c6c\": container with ID starting with 98c3a84bb3719670b27f639d1ea40727f5a8a516d9dc7b62648e43e0b2b56c6c not found: ID does not exist" containerID="98c3a84bb3719670b27f639d1ea40727f5a8a516d9dc7b62648e43e0b2b56c6c" Dec 06 16:53:51 crc kubenswrapper[4813]: I1206 16:53:51.038838 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98c3a84bb3719670b27f639d1ea40727f5a8a516d9dc7b62648e43e0b2b56c6c"} err="failed to get container status \"98c3a84bb3719670b27f639d1ea40727f5a8a516d9dc7b62648e43e0b2b56c6c\": rpc error: code = NotFound desc = could not find container \"98c3a84bb3719670b27f639d1ea40727f5a8a516d9dc7b62648e43e0b2b56c6c\": container with ID starting with 98c3a84bb3719670b27f639d1ea40727f5a8a516d9dc7b62648e43e0b2b56c6c not found: ID does not exist" Dec 06 16:53:51 crc kubenswrapper[4813]: I1206 16:53:51.038867 4813 scope.go:117] "RemoveContainer" containerID="b87b9060f94bb5e5d2af15a74f6ac48e12bb89c8c9cb8e37f4d1431950a367d8" Dec 06 16:53:51 crc kubenswrapper[4813]: E1206 16:53:51.039153 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b87b9060f94bb5e5d2af15a74f6ac48e12bb89c8c9cb8e37f4d1431950a367d8\": container with ID starting with b87b9060f94bb5e5d2af15a74f6ac48e12bb89c8c9cb8e37f4d1431950a367d8 not found: ID does not exist" containerID="b87b9060f94bb5e5d2af15a74f6ac48e12bb89c8c9cb8e37f4d1431950a367d8" Dec 06 16:53:51 crc kubenswrapper[4813]: I1206 16:53:51.039241 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b87b9060f94bb5e5d2af15a74f6ac48e12bb89c8c9cb8e37f4d1431950a367d8"} err="failed to get container status \"b87b9060f94bb5e5d2af15a74f6ac48e12bb89c8c9cb8e37f4d1431950a367d8\": rpc error: code = NotFound desc = could not find container \"b87b9060f94bb5e5d2af15a74f6ac48e12bb89c8c9cb8e37f4d1431950a367d8\": container with ID starting with b87b9060f94bb5e5d2af15a74f6ac48e12bb89c8c9cb8e37f4d1431950a367d8 not found: ID does not exist" Dec 06 16:53:52 crc kubenswrapper[4813]: I1206 16:53:52.496907 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e5505e6-cb46-4c1a-81d6-3f4d970291e7" path="/var/lib/kubelet/pods/8e5505e6-cb46-4c1a-81d6-3f4d970291e7/volumes" Dec 06 16:53:53 crc kubenswrapper[4813]: I1206 16:53:53.433982 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jgn6l/must-gather-jb2rq"] Dec 06 16:53:53 crc kubenswrapper[4813]: I1206 16:53:53.434610 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-jgn6l/must-gather-jb2rq" podUID="1a835870-e23a-446e-8c38-fe587e9cee6b" containerName="copy" containerID="cri-o://e3d87520cb39ce4ee89b161ddfb351a6c242a2718c3f771de3ebdafd24aec70e" gracePeriod=2 Dec 06 16:53:53 crc kubenswrapper[4813]: I1206 16:53:53.453683 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jgn6l/must-gather-jb2rq"] Dec 06 16:53:53 crc kubenswrapper[4813]: I1206 16:53:53.863545 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jgn6l_must-gather-jb2rq_1a835870-e23a-446e-8c38-fe587e9cee6b/copy/0.log" Dec 06 16:53:53 crc kubenswrapper[4813]: I1206 16:53:53.864119 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgn6l/must-gather-jb2rq" Dec 06 16:53:53 crc kubenswrapper[4813]: I1206 16:53:53.990002 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1a835870-e23a-446e-8c38-fe587e9cee6b-must-gather-output\") pod \"1a835870-e23a-446e-8c38-fe587e9cee6b\" (UID: \"1a835870-e23a-446e-8c38-fe587e9cee6b\") " Dec 06 16:53:53 crc kubenswrapper[4813]: I1206 16:53:53.990419 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndcqn\" (UniqueName: \"kubernetes.io/projected/1a835870-e23a-446e-8c38-fe587e9cee6b-kube-api-access-ndcqn\") pod \"1a835870-e23a-446e-8c38-fe587e9cee6b\" (UID: \"1a835870-e23a-446e-8c38-fe587e9cee6b\") " Dec 06 16:53:53 crc kubenswrapper[4813]: I1206 16:53:53.991668 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jgn6l_must-gather-jb2rq_1a835870-e23a-446e-8c38-fe587e9cee6b/copy/0.log" Dec 06 16:53:53 crc kubenswrapper[4813]: I1206 16:53:53.991985 4813 generic.go:334] "Generic (PLEG): container finished" podID="1a835870-e23a-446e-8c38-fe587e9cee6b" containerID="e3d87520cb39ce4ee89b161ddfb351a6c242a2718c3f771de3ebdafd24aec70e" exitCode=143 Dec 06 16:53:53 crc kubenswrapper[4813]: I1206 16:53:53.992038 4813 scope.go:117] "RemoveContainer" containerID="e3d87520cb39ce4ee89b161ddfb351a6c242a2718c3f771de3ebdafd24aec70e" Dec 06 16:53:53 crc kubenswrapper[4813]: I1206 16:53:53.992149 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jgn6l/must-gather-jb2rq" Dec 06 16:53:54 crc kubenswrapper[4813]: I1206 16:53:53.997557 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a835870-e23a-446e-8c38-fe587e9cee6b-kube-api-access-ndcqn" (OuterVolumeSpecName: "kube-api-access-ndcqn") pod "1a835870-e23a-446e-8c38-fe587e9cee6b" (UID: "1a835870-e23a-446e-8c38-fe587e9cee6b"). InnerVolumeSpecName "kube-api-access-ndcqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:53:54 crc kubenswrapper[4813]: I1206 16:53:54.042809 4813 scope.go:117] "RemoveContainer" containerID="db9b299109c15116a10e92800ddbe7711c0fbc10c48090abc39a6e10048f6867" Dec 06 16:53:54 crc kubenswrapper[4813]: I1206 16:53:54.090059 4813 scope.go:117] "RemoveContainer" containerID="e3d87520cb39ce4ee89b161ddfb351a6c242a2718c3f771de3ebdafd24aec70e" Dec 06 16:53:54 crc kubenswrapper[4813]: E1206 16:53:54.090461 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3d87520cb39ce4ee89b161ddfb351a6c242a2718c3f771de3ebdafd24aec70e\": container with ID starting with e3d87520cb39ce4ee89b161ddfb351a6c242a2718c3f771de3ebdafd24aec70e not found: ID does not exist" containerID="e3d87520cb39ce4ee89b161ddfb351a6c242a2718c3f771de3ebdafd24aec70e" Dec 06 16:53:54 crc kubenswrapper[4813]: I1206 16:53:54.090502 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3d87520cb39ce4ee89b161ddfb351a6c242a2718c3f771de3ebdafd24aec70e"} err="failed to get container status \"e3d87520cb39ce4ee89b161ddfb351a6c242a2718c3f771de3ebdafd24aec70e\": rpc error: code = NotFound desc = could not find container \"e3d87520cb39ce4ee89b161ddfb351a6c242a2718c3f771de3ebdafd24aec70e\": container with ID starting with e3d87520cb39ce4ee89b161ddfb351a6c242a2718c3f771de3ebdafd24aec70e not found: ID does not exist" Dec 06 16:53:54 crc kubenswrapper[4813]: I1206 16:53:54.090550 4813 scope.go:117] "RemoveContainer" containerID="db9b299109c15116a10e92800ddbe7711c0fbc10c48090abc39a6e10048f6867" Dec 06 16:53:54 crc kubenswrapper[4813]: E1206 16:53:54.090815 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db9b299109c15116a10e92800ddbe7711c0fbc10c48090abc39a6e10048f6867\": container with ID starting with db9b299109c15116a10e92800ddbe7711c0fbc10c48090abc39a6e10048f6867 not found: ID does not exist" containerID="db9b299109c15116a10e92800ddbe7711c0fbc10c48090abc39a6e10048f6867" Dec 06 16:53:54 crc kubenswrapper[4813]: I1206 16:53:54.090838 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db9b299109c15116a10e92800ddbe7711c0fbc10c48090abc39a6e10048f6867"} err="failed to get container status \"db9b299109c15116a10e92800ddbe7711c0fbc10c48090abc39a6e10048f6867\": rpc error: code = NotFound desc = could not find container \"db9b299109c15116a10e92800ddbe7711c0fbc10c48090abc39a6e10048f6867\": container with ID starting with db9b299109c15116a10e92800ddbe7711c0fbc10c48090abc39a6e10048f6867 not found: ID does not exist" Dec 06 16:53:54 crc kubenswrapper[4813]: I1206 16:53:54.092194 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndcqn\" (UniqueName: \"kubernetes.io/projected/1a835870-e23a-446e-8c38-fe587e9cee6b-kube-api-access-ndcqn\") on node \"crc\" DevicePath \"\"" Dec 06 16:53:54 crc kubenswrapper[4813]: I1206 16:53:54.175587 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a835870-e23a-446e-8c38-fe587e9cee6b-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "1a835870-e23a-446e-8c38-fe587e9cee6b" (UID: "1a835870-e23a-446e-8c38-fe587e9cee6b"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:53:54 crc kubenswrapper[4813]: I1206 16:53:54.194362 4813 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1a835870-e23a-446e-8c38-fe587e9cee6b-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 06 16:53:54 crc kubenswrapper[4813]: I1206 16:53:54.496052 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a835870-e23a-446e-8c38-fe587e9cee6b" path="/var/lib/kubelet/pods/1a835870-e23a-446e-8c38-fe587e9cee6b/volumes" Dec 06 16:54:19 crc kubenswrapper[4813]: I1206 16:54:19.427914 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:54:19 crc kubenswrapper[4813]: I1206 16:54:19.428680 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:54:49 crc kubenswrapper[4813]: I1206 16:54:49.427621 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:54:49 crc kubenswrapper[4813]: I1206 16:54:49.428396 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:55:19 crc kubenswrapper[4813]: I1206 16:55:19.428081 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:55:19 crc kubenswrapper[4813]: I1206 16:55:19.428737 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:55:19 crc kubenswrapper[4813]: I1206 16:55:19.428831 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" Dec 06 16:55:19 crc kubenswrapper[4813]: I1206 16:55:19.429866 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fcb495a2496cefa1a08b361825efb708538fc1c2d99f9878beb2279f9088eca6"} pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 16:55:19 crc kubenswrapper[4813]: I1206 16:55:19.429947 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" containerID="cri-o://fcb495a2496cefa1a08b361825efb708538fc1c2d99f9878beb2279f9088eca6" gracePeriod=600 Dec 06 16:55:19 crc kubenswrapper[4813]: I1206 16:55:19.956840 4813 generic.go:334] "Generic (PLEG): container finished" podID="d88e8bae-c055-4c55-b548-f621ff96de06" containerID="fcb495a2496cefa1a08b361825efb708538fc1c2d99f9878beb2279f9088eca6" exitCode=0 Dec 06 16:55:19 crc kubenswrapper[4813]: I1206 16:55:19.956875 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerDied","Data":"fcb495a2496cefa1a08b361825efb708538fc1c2d99f9878beb2279f9088eca6"} Dec 06 16:55:19 crc kubenswrapper[4813]: I1206 16:55:19.957543 4813 scope.go:117] "RemoveContainer" containerID="969c7429b9faa2222909b4bad4ef827351b31f2cde6d6ab68e5af156f1ab0dfc" Dec 06 16:55:20 crc kubenswrapper[4813]: I1206 16:55:20.972915 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" event={"ID":"d88e8bae-c055-4c55-b548-f621ff96de06","Type":"ContainerStarted","Data":"e0830d30ee35cbe6d389ca1a318bf782011e79403c5eaa194b7817e2b8741c93"} Dec 06 16:55:32 crc kubenswrapper[4813]: I1206 16:55:32.500501 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-68zbh"] Dec 06 16:55:32 crc kubenswrapper[4813]: E1206 16:55:32.501894 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a835870-e23a-446e-8c38-fe587e9cee6b" containerName="gather" Dec 06 16:55:32 crc kubenswrapper[4813]: I1206 16:55:32.501918 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a835870-e23a-446e-8c38-fe587e9cee6b" containerName="gather" Dec 06 16:55:32 crc kubenswrapper[4813]: E1206 16:55:32.501941 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a835870-e23a-446e-8c38-fe587e9cee6b" containerName="copy" Dec 06 16:55:32 crc kubenswrapper[4813]: I1206 16:55:32.501953 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a835870-e23a-446e-8c38-fe587e9cee6b" containerName="copy" Dec 06 16:55:32 crc kubenswrapper[4813]: E1206 16:55:32.501970 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e5505e6-cb46-4c1a-81d6-3f4d970291e7" containerName="registry-server" Dec 06 16:55:32 crc kubenswrapper[4813]: I1206 16:55:32.501982 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e5505e6-cb46-4c1a-81d6-3f4d970291e7" containerName="registry-server" Dec 06 16:55:32 crc kubenswrapper[4813]: E1206 16:55:32.502028 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e5505e6-cb46-4c1a-81d6-3f4d970291e7" containerName="extract-utilities" Dec 06 16:55:32 crc kubenswrapper[4813]: I1206 16:55:32.502040 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e5505e6-cb46-4c1a-81d6-3f4d970291e7" containerName="extract-utilities" Dec 06 16:55:32 crc kubenswrapper[4813]: E1206 16:55:32.502095 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e5505e6-cb46-4c1a-81d6-3f4d970291e7" containerName="extract-content" Dec 06 16:55:32 crc kubenswrapper[4813]: I1206 16:55:32.502107 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e5505e6-cb46-4c1a-81d6-3f4d970291e7" containerName="extract-content" Dec 06 16:55:32 crc kubenswrapper[4813]: I1206 16:55:32.502499 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a835870-e23a-446e-8c38-fe587e9cee6b" containerName="copy" Dec 06 16:55:32 crc kubenswrapper[4813]: I1206 16:55:32.502545 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a835870-e23a-446e-8c38-fe587e9cee6b" containerName="gather" Dec 06 16:55:32 crc kubenswrapper[4813]: I1206 16:55:32.502575 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e5505e6-cb46-4c1a-81d6-3f4d970291e7" containerName="registry-server" Dec 06 16:55:32 crc kubenswrapper[4813]: I1206 16:55:32.505018 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-68zbh" Dec 06 16:55:32 crc kubenswrapper[4813]: I1206 16:55:32.510383 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-68zbh"] Dec 06 16:55:32 crc kubenswrapper[4813]: I1206 16:55:32.606693 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cea513c-9f57-43a2-b614-87a829a596b6-catalog-content\") pod \"redhat-marketplace-68zbh\" (UID: \"9cea513c-9f57-43a2-b614-87a829a596b6\") " pod="openshift-marketplace/redhat-marketplace-68zbh" Dec 06 16:55:32 crc kubenswrapper[4813]: I1206 16:55:32.606760 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cea513c-9f57-43a2-b614-87a829a596b6-utilities\") pod \"redhat-marketplace-68zbh\" (UID: \"9cea513c-9f57-43a2-b614-87a829a596b6\") " pod="openshift-marketplace/redhat-marketplace-68zbh" Dec 06 16:55:32 crc kubenswrapper[4813]: I1206 16:55:32.606845 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cn279\" (UniqueName: \"kubernetes.io/projected/9cea513c-9f57-43a2-b614-87a829a596b6-kube-api-access-cn279\") pod \"redhat-marketplace-68zbh\" (UID: \"9cea513c-9f57-43a2-b614-87a829a596b6\") " pod="openshift-marketplace/redhat-marketplace-68zbh" Dec 06 16:55:32 crc kubenswrapper[4813]: I1206 16:55:32.708626 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cea513c-9f57-43a2-b614-87a829a596b6-catalog-content\") pod \"redhat-marketplace-68zbh\" (UID: \"9cea513c-9f57-43a2-b614-87a829a596b6\") " pod="openshift-marketplace/redhat-marketplace-68zbh" Dec 06 16:55:32 crc kubenswrapper[4813]: I1206 16:55:32.708688 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cea513c-9f57-43a2-b614-87a829a596b6-utilities\") pod \"redhat-marketplace-68zbh\" (UID: \"9cea513c-9f57-43a2-b614-87a829a596b6\") " pod="openshift-marketplace/redhat-marketplace-68zbh" Dec 06 16:55:32 crc kubenswrapper[4813]: I1206 16:55:32.708737 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cn279\" (UniqueName: \"kubernetes.io/projected/9cea513c-9f57-43a2-b614-87a829a596b6-kube-api-access-cn279\") pod \"redhat-marketplace-68zbh\" (UID: \"9cea513c-9f57-43a2-b614-87a829a596b6\") " pod="openshift-marketplace/redhat-marketplace-68zbh" Dec 06 16:55:32 crc kubenswrapper[4813]: I1206 16:55:32.709355 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cea513c-9f57-43a2-b614-87a829a596b6-catalog-content\") pod \"redhat-marketplace-68zbh\" (UID: \"9cea513c-9f57-43a2-b614-87a829a596b6\") " pod="openshift-marketplace/redhat-marketplace-68zbh" Dec 06 16:55:32 crc kubenswrapper[4813]: I1206 16:55:32.709363 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cea513c-9f57-43a2-b614-87a829a596b6-utilities\") pod \"redhat-marketplace-68zbh\" (UID: \"9cea513c-9f57-43a2-b614-87a829a596b6\") " pod="openshift-marketplace/redhat-marketplace-68zbh" Dec 06 16:55:32 crc kubenswrapper[4813]: I1206 16:55:32.733568 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cn279\" (UniqueName: \"kubernetes.io/projected/9cea513c-9f57-43a2-b614-87a829a596b6-kube-api-access-cn279\") pod \"redhat-marketplace-68zbh\" (UID: \"9cea513c-9f57-43a2-b614-87a829a596b6\") " pod="openshift-marketplace/redhat-marketplace-68zbh" Dec 06 16:55:32 crc kubenswrapper[4813]: I1206 16:55:32.828925 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-68zbh" Dec 06 16:55:33 crc kubenswrapper[4813]: I1206 16:55:33.284077 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-68zbh"] Dec 06 16:55:33 crc kubenswrapper[4813]: W1206 16:55:33.298858 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9cea513c_9f57_43a2_b614_87a829a596b6.slice/crio-c0260bf01bb292ecbe9d4e962d289308c7ef17c575644e4393973665a066b6fb WatchSource:0}: Error finding container c0260bf01bb292ecbe9d4e962d289308c7ef17c575644e4393973665a066b6fb: Status 404 returned error can't find the container with id c0260bf01bb292ecbe9d4e962d289308c7ef17c575644e4393973665a066b6fb Dec 06 16:55:34 crc kubenswrapper[4813]: I1206 16:55:34.111042 4813 generic.go:334] "Generic (PLEG): container finished" podID="9cea513c-9f57-43a2-b614-87a829a596b6" containerID="44e50f45e5672b36183cff5e2dd5a8ed4baa68befd3981e539bcc6146d2b89ba" exitCode=0 Dec 06 16:55:34 crc kubenswrapper[4813]: I1206 16:55:34.111156 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68zbh" event={"ID":"9cea513c-9f57-43a2-b614-87a829a596b6","Type":"ContainerDied","Data":"44e50f45e5672b36183cff5e2dd5a8ed4baa68befd3981e539bcc6146d2b89ba"} Dec 06 16:55:34 crc kubenswrapper[4813]: I1206 16:55:34.111438 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68zbh" event={"ID":"9cea513c-9f57-43a2-b614-87a829a596b6","Type":"ContainerStarted","Data":"c0260bf01bb292ecbe9d4e962d289308c7ef17c575644e4393973665a066b6fb"} Dec 06 16:55:35 crc kubenswrapper[4813]: I1206 16:55:35.121432 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68zbh" event={"ID":"9cea513c-9f57-43a2-b614-87a829a596b6","Type":"ContainerStarted","Data":"12b7899828cf9e4c92b7d45c64c578bd46cabd8e1a216f15b263d0b0dffa1efe"} Dec 06 16:55:36 crc kubenswrapper[4813]: I1206 16:55:36.135864 4813 generic.go:334] "Generic (PLEG): container finished" podID="9cea513c-9f57-43a2-b614-87a829a596b6" containerID="12b7899828cf9e4c92b7d45c64c578bd46cabd8e1a216f15b263d0b0dffa1efe" exitCode=0 Dec 06 16:55:36 crc kubenswrapper[4813]: I1206 16:55:36.136027 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68zbh" event={"ID":"9cea513c-9f57-43a2-b614-87a829a596b6","Type":"ContainerDied","Data":"12b7899828cf9e4c92b7d45c64c578bd46cabd8e1a216f15b263d0b0dffa1efe"} Dec 06 16:55:37 crc kubenswrapper[4813]: I1206 16:55:37.154444 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68zbh" event={"ID":"9cea513c-9f57-43a2-b614-87a829a596b6","Type":"ContainerStarted","Data":"2600322740f12cb0c943fe836c2464292cfeaed32ccc6eca727921dcc5911141"} Dec 06 16:55:42 crc kubenswrapper[4813]: I1206 16:55:42.829042 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-68zbh" Dec 06 16:55:42 crc kubenswrapper[4813]: I1206 16:55:42.829484 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-68zbh" Dec 06 16:55:42 crc kubenswrapper[4813]: I1206 16:55:42.903639 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-68zbh" Dec 06 16:55:42 crc kubenswrapper[4813]: I1206 16:55:42.932222 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-68zbh" podStartSLOduration=8.513480319 podStartE2EDuration="10.932204987s" podCreationTimestamp="2025-12-06 16:55:32 +0000 UTC" firstStartedPulling="2025-12-06 16:55:34.116426526 +0000 UTC m=+4174.007306112" lastFinishedPulling="2025-12-06 16:55:36.535151164 +0000 UTC m=+4176.426030780" observedRunningTime="2025-12-06 16:55:37.181395018 +0000 UTC m=+4177.072274634" watchObservedRunningTime="2025-12-06 16:55:42.932204987 +0000 UTC m=+4182.823084583" Dec 06 16:55:43 crc kubenswrapper[4813]: I1206 16:55:43.965577 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-68zbh" Dec 06 16:55:47 crc kubenswrapper[4813]: I1206 16:55:47.555079 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-68zbh"] Dec 06 16:55:47 crc kubenswrapper[4813]: I1206 16:55:47.555901 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-68zbh" podUID="9cea513c-9f57-43a2-b614-87a829a596b6" containerName="registry-server" containerID="cri-o://2600322740f12cb0c943fe836c2464292cfeaed32ccc6eca727921dcc5911141" gracePeriod=2 Dec 06 16:55:48 crc kubenswrapper[4813]: I1206 16:55:48.345654 4813 generic.go:334] "Generic (PLEG): container finished" podID="9cea513c-9f57-43a2-b614-87a829a596b6" containerID="2600322740f12cb0c943fe836c2464292cfeaed32ccc6eca727921dcc5911141" exitCode=0 Dec 06 16:55:48 crc kubenswrapper[4813]: I1206 16:55:48.345759 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68zbh" event={"ID":"9cea513c-9f57-43a2-b614-87a829a596b6","Type":"ContainerDied","Data":"2600322740f12cb0c943fe836c2464292cfeaed32ccc6eca727921dcc5911141"} Dec 06 16:55:48 crc kubenswrapper[4813]: I1206 16:55:48.595349 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-68zbh" Dec 06 16:55:48 crc kubenswrapper[4813]: I1206 16:55:48.677481 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cea513c-9f57-43a2-b614-87a829a596b6-catalog-content\") pod \"9cea513c-9f57-43a2-b614-87a829a596b6\" (UID: \"9cea513c-9f57-43a2-b614-87a829a596b6\") " Dec 06 16:55:48 crc kubenswrapper[4813]: I1206 16:55:48.677585 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cea513c-9f57-43a2-b614-87a829a596b6-utilities\") pod \"9cea513c-9f57-43a2-b614-87a829a596b6\" (UID: \"9cea513c-9f57-43a2-b614-87a829a596b6\") " Dec 06 16:55:48 crc kubenswrapper[4813]: I1206 16:55:48.677615 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cn279\" (UniqueName: \"kubernetes.io/projected/9cea513c-9f57-43a2-b614-87a829a596b6-kube-api-access-cn279\") pod \"9cea513c-9f57-43a2-b614-87a829a596b6\" (UID: \"9cea513c-9f57-43a2-b614-87a829a596b6\") " Dec 06 16:55:48 crc kubenswrapper[4813]: I1206 16:55:48.678796 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cea513c-9f57-43a2-b614-87a829a596b6-utilities" (OuterVolumeSpecName: "utilities") pod "9cea513c-9f57-43a2-b614-87a829a596b6" (UID: "9cea513c-9f57-43a2-b614-87a829a596b6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:55:48 crc kubenswrapper[4813]: I1206 16:55:48.692964 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cea513c-9f57-43a2-b614-87a829a596b6-kube-api-access-cn279" (OuterVolumeSpecName: "kube-api-access-cn279") pod "9cea513c-9f57-43a2-b614-87a829a596b6" (UID: "9cea513c-9f57-43a2-b614-87a829a596b6"). InnerVolumeSpecName "kube-api-access-cn279". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 16:55:48 crc kubenswrapper[4813]: I1206 16:55:48.701400 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cea513c-9f57-43a2-b614-87a829a596b6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9cea513c-9f57-43a2-b614-87a829a596b6" (UID: "9cea513c-9f57-43a2-b614-87a829a596b6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 16:55:48 crc kubenswrapper[4813]: I1206 16:55:48.779820 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cea513c-9f57-43a2-b614-87a829a596b6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 16:55:48 crc kubenswrapper[4813]: I1206 16:55:48.780308 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cea513c-9f57-43a2-b614-87a829a596b6-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 16:55:48 crc kubenswrapper[4813]: I1206 16:55:48.780332 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cn279\" (UniqueName: \"kubernetes.io/projected/9cea513c-9f57-43a2-b614-87a829a596b6-kube-api-access-cn279\") on node \"crc\" DevicePath \"\"" Dec 06 16:55:49 crc kubenswrapper[4813]: I1206 16:55:49.362422 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68zbh" event={"ID":"9cea513c-9f57-43a2-b614-87a829a596b6","Type":"ContainerDied","Data":"c0260bf01bb292ecbe9d4e962d289308c7ef17c575644e4393973665a066b6fb"} Dec 06 16:55:49 crc kubenswrapper[4813]: I1206 16:55:49.362482 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-68zbh" Dec 06 16:55:49 crc kubenswrapper[4813]: I1206 16:55:49.362542 4813 scope.go:117] "RemoveContainer" containerID="2600322740f12cb0c943fe836c2464292cfeaed32ccc6eca727921dcc5911141" Dec 06 16:55:49 crc kubenswrapper[4813]: I1206 16:55:49.403707 4813 scope.go:117] "RemoveContainer" containerID="12b7899828cf9e4c92b7d45c64c578bd46cabd8e1a216f15b263d0b0dffa1efe" Dec 06 16:55:49 crc kubenswrapper[4813]: I1206 16:55:49.416334 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-68zbh"] Dec 06 16:55:49 crc kubenswrapper[4813]: I1206 16:55:49.443379 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-68zbh"] Dec 06 16:55:49 crc kubenswrapper[4813]: I1206 16:55:49.444188 4813 scope.go:117] "RemoveContainer" containerID="44e50f45e5672b36183cff5e2dd5a8ed4baa68befd3981e539bcc6146d2b89ba" Dec 06 16:55:50 crc kubenswrapper[4813]: I1206 16:55:50.503956 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cea513c-9f57-43a2-b614-87a829a596b6" path="/var/lib/kubelet/pods/9cea513c-9f57-43a2-b614-87a829a596b6/volumes" Dec 06 16:57:49 crc kubenswrapper[4813]: I1206 16:57:49.428590 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:57:49 crc kubenswrapper[4813]: I1206 16:57:49.429327 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 16:58:19 crc kubenswrapper[4813]: I1206 16:58:19.428257 4813 patch_prober.go:28] interesting pod/machine-config-daemon-t5xp8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 16:58:19 crc kubenswrapper[4813]: I1206 16:58:19.428926 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t5xp8" podUID="d88e8bae-c055-4c55-b548-f621ff96de06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515115060262024443 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015115060263017361 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015115047317016511 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015115047317015461 5ustar corecore